Skip Navigation
InitialsDiceBearhttps://github.com/dicebear/dicebearhttps://creativecommons.org/publicdomain/zero/1.0/„Initials” (https://github.com/dicebear/dicebear) by „DiceBear”, licensed under „CC0 1.0” (https://creativecommons.org/publicdomain/zero/1.0/)IN
Posts
58
Comments
131
Joined
2 yr. ago
ShareGPT @lemmy.fmhy.ml
InternetPirate @lemmy.fmhy.ml

Tingling and palpitations spreading throughout the body

The tingling and palpitations that have been spreading throughout the body can be caused by several conditions. Here are some possibilities based on the search results:

  • Episodic Neurologic Symptoms: Patients with this condition usually complain of severe apprehension, dizziness, chest discomfort, palpitations, or fear of "losing control." The onset is usually under age 40[1].
  • Heart Attack or Panic Attack: Both conditions can cause chest pain, shortness of breath, dizziness, sweating, passing out, tingling, or a sensation of impending doom. These shared symptoms of heart and panic attack may also be caused by other serious conditions such as blood clots, lung infection or collapse, or tear in the large vessels of the chest for patients with certain pre-existing risk factors[2].
  • Atrial Fibrillation: The most obvious symptom of this condition is heart palpitations, where the heart feels like it's pounding, fluttering, or beating irregularly, often for a few seconds or
Singularity | Artificial Intelligence (ai), Technology & Futurology @lemmy.fmhy.ml
InternetPirate @lemmy.fmhy.ml

China Issues Rules for Generative AI, Mandating Adherence to 'Socialist Values' (article from 14.07.2023)

  • There are several Linux command-line tools that can be used to download a mirror of a website for offline browsing. Here are some of the most popular ones:

    1. wget: wget is a free utility for non-interactive download of files from the Web. It supports HTTP, HTTPS, and FTP protocols, as well as retrieval through HTTP proxies. To download a website recursively, use the -r option. For example, the following command will download the entire website located at http://example.com:
       undefined
          
      wget --mirror --convert-links --adjust-extension --page-requisites --no-parent http://example.com
      
        
      This will create a local copy of the website in a directory named example.com.
    2. httrack: httrack is a free and open-source website copier that allows you to download a website and browse it offline. It supports HTTP, HTTPS, and FTP protocols, as well as retrieval through HTTP proxies. To download a website, use the -O option followed by the directory where you want to save the website. For example, the following command will download the entire website located at http://example.com:
       undefined
          
      httrack http://example.com -O /path/to/save/directory
      
        
      This will create a local copy of the website in the directory /path/to/save/directory/example.com.
    3. curl: curl is a command-line tool for transferring data from or to a server, using one of the supported protocols (HTTP, HTTPS, FTP, etc.). To download a website recursively, use the -r option. For example, the following command will download the entire website located at http://example.com:
       undefined
          
      curl -k -L -O -J -R -H 'Referer: http://example.com' -e 'http://example.com' -A 'Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/58.0.3029.110 Safari/537.36' -m 1800 --retry 3 --retry-delay 5 --retry-max-time 60 --retry-connrefused -s -S -L -r -k -J -O -C - http://example.com
      
        
      This will create a local copy of the website in the current directory.

    These tools have different options and features, so you should choose the one that best suits your needs.

    Citations: [1] https://www.tutorialspoint.com/5-linux-command-line-based-tools-for-downloading-files-and-browsing-websites [2] https://www.kali.org/tools/httrack/ [3] https://medevel.com/os1-3-web-copier/ [4] http://www.linux-magazine.com/Online/Features/WebHTTrack-Website-Copier [5] https://winaero.com/make-offline-copy-of-a-site-with-wget-on-windows-and-linux/ [6] https://alvinalexander.com/linux-unix/how-to-make-offline-mirror-copy-website-with-wget

  • wget -mkEpnp

    wget --mirror --convert-links --adjust-extension --page-requisites --no-parent http://example.org

    Explanation of the various flags:

     undefined
            --mirror – Makes (among other things) the download recursive.
        --convert-links – convert all the links (also to stuff like CSS stylesheets) to relative, so it will be suitable for offline viewing.
        --adjust-extension – Adds suitable extensions to filenames (html or css) depending on their content-type.
        --page-requisites – Download things like CSS style-sheets and images required to properly display the page offline.
        --no-parent – When recursing do not ascend to the parent directory. It useful for restricting the download to only a portion of the site.
    
    
      

    wget -mpHkKEb -t 1 -e robots=off -U 'Mozilla/5.0 (X11; Ubuntu; Linux x86_64; rv:40.0) Gecko/20100101 Firefox/40.0' http://www.example.com

     undefined
            –m (--mirror) : turn on options suitable for mirroring (infinite recursive download and timestamps).
    
        -p (--page-requisites) : download all files that are necessary to properly display a given HTML page. This includes such things as inlined images, sounds, and referenced stylesheets.
    
        -H (--span-hosts): enable spanning across hosts when doing recursive retrieving.
    
        –k (--convert-links) : after the download, convert the links in document for local viewing.
    
        -K (--backup-converted) : when converting a file, back up the original version with a .orig suffix. Affects the behavior of -N.
    
        -E (--adjust-extension) : add the proper extension to the end of the file.
    
        -b (--background) : go to background immediately after startup. If no output file is specified via the -o, output is redirected to wget-log.
    
        -e (--execute) : execute command (robots=off).
    
        -t number (--tries=number) : set number of tries to number.
    
        -U (--user-agent) : identify as agent-string to the HTTP server. Some servers may ban you permanently for recursively download if you send the default User Agent.
    
    
      

    Cronjobs

    0 23 * * * cd ~/Documents/Webs/mirror; wget -mpk -t 1 -e robots=off -U 'Mozilla/5.0 (X11; Ubuntu; Linux x86_64; rv:40.0) Gecko/20100101 Firefox/40.0' https://example.com

    0 23 * * * cd ~/Documents/Webs/mirror; wget -mpk -t 1 -e robots=off -U 'Mozilla/5.0 (X11; Ubuntu; Linux x86_64; rv:40.0) Gecko/20100101 Firefox/40.0' https://example.com

    0 23 * * * cd ~/Documents/Webs/mirror; wget -mpk t 1 -e robots=off -U 'Mozilla/5.0 (X11; Ubuntu; Linux x86_64; rv:40.0) Gecko/20100101 Firefox/40.0' https://example.com

    0 23 * * * cd ~/Documents/Webs/mirror; wget -mpkH t 1 -e robots=off -U 'Mozilla/5.0 (X11; Ubuntu; Linux x86_64; rv:40.0) Gecko/20100101 Firefox/40.0' -D https://example.com

    0 23 * * * cd ~/Documents/Webs/mirror; wget -mpk t 1 -e robots=off -U 'Mozilla/5.0 (X11; Ubuntu; Linux x86_64; rv:40.0) Gecko/20100101 Firefox/40.0' https://example.com

    0 23 * * * cd ~/Documents/Webs/mirror; wget -mpk t 1 -e robots=off -U 'Mozilla/5.0 (X11; Ubuntu; Linux x86_64; rv:40.0) Gecko/20100101 Firefox/40.0' https://example.com

    0 23 * 1 * cd ~/Documents/Webs/mirror; wget -mpk t 1 -e robots=off -U 'Mozilla/5.0 (X11; Ubuntu; Linux x86_64; rv:40.0) Gecko/20100101 Firefox/40.0' https://example.com

    0 8 * * * pkill wget; cd ~/Documents/Webs/mirror/; rm wget*

  • ShareGPT @lemmy.fmhy.ml
    InternetPirate @lemmy.fmhy.ml

    How to download a website for offline browsing using wget?

    To download a website for offline browsing using wget, you can use the following command:

     bash
        
    wget --recursive --no-clobber --page-requisites --html-extension --convert-links --restrict-file-names=windows --domains example.com --no-parent http://www.example.com
    
      

    Replace example.com with the website you want to download. This command will download the entire website, including all the elements that compose the page (images, CSS, etc.), and convert the links so that they work locally and offline[1].

    Here's a brief explanation of the options used in the command:

    • --recursive: Download the entire website recursively.
    • --no-clobber: Don't overwrite existing files.
    • --page-requisites: Download all the elements required to display the page properly (images, CSS, etc.).
    • --html-extension: Save files with the .html extension.
    • --convert-links: Convert links so that they work locally and offline.
    • --restrict-file-names=windows: Modify filenames so that they wor
    Singularity | Artificial Intelligence (ai), Technology & Futurology @lemmy.fmhy.ml
    InternetPirate @lemmy.fmhy.ml

    Mind blown 🤯🤯

    Bing (multimodal) image input is free!

    Lemmy @lemmy.ml
    InternetPirate @lemmy.fmhy.ml

    Post title character limit is 201

    I couldn't find this using the search.

  • To achieve a rate of 1.39 requests per second, you can use a global variable to keep track of the time elapsed between requests and then calculate the delay based on that time. Here's a modified version of your fetch_github_data function that implements this:

     python
        
    import time
    import requests
    import logging
    
    PERSONAL_ACCESS_TOKEN = "your_personal_access_token"
    DELAY = 1 / 1.39  # Calculate the delay for 1.39 requests per second
    last_request_time = 0  # Initialize a global variable to store the last request time
    
    
    def fetch_github_data(url):
        global last_request_time  # Access the global variable
    
        try:
            headers = {
                "Accept": "application/vnd.github+json",
                "Authorization": f"Bearer {PERSONAL_ACCESS_TOKEN}",
                "X-GitHub-Api-Version": "2022-11-28",
            }
    
            # Calculate the time elapsed since the last request
            time_elapsed = time.time() - last_request_time
    
            # Calculate the required delay based on the time elapsed
            required_delay = max(0, DELAY - time_elapsed)
    
            # Sleep for the required delay
            time.sleep(required_delay)
    
            response = requests.get(url, headers=headers)
    
            # Update the last request time
            last_request_time = time.time()
    
            logging.info(f"Fetched data from {url}")
            return response.json()
        except requests.exceptions.RequestException as e:
            logging.exception(f"Error fetching data from {url}\n{e}")
            raise
    
      

    This code calculates the required delay based on the desired rate of 1.39 requests per second and the time elapsed since the last request. It then sleeps for the required delay before making the next request. The global variable last_request_time is used to keep track of the time of the last request.

    Citations: [1] https://www.geeksforgeeks.org/how-to-add-time-delay-in-python/ [2] https://stackoverflow.com/questions/66229987/calculate-attempts-per-second [3] https://pypi.org/project/requests-ratelimiter/ [4] https://akshayranganath.github.io/Rate-Limiting-With-Python/ [5] https://stackoverflow.com/questions/32815451/are-global-variables-thread-safe-in-flask-how-do-i-share-data-between-requests [6] https://stackoverflow.com/questions/44014718/python-request-get-after-few-seconds [7] https://realpython.com/python-sleep/ [8] https://algotrading101.com/learn/yahoo-finance-api-guide/ [9] https://stackoverflow.com/questions/26098711/limiting-number-of-http-requests-per-second-on-python [10] https://realpython.com/python-use-global-variable-in-function/ [11] https://scrapeops.io/python-scrapy-playbook/scrapy-delay-between-requests/ [12] https://cloud.google.com/python/docs/reference/storage/1.44.0/client [13] https://github.com/JWCook/requests-ratelimiter [14] https://discuss.python.org/t/global-variables-shared-across-modules/16833 [15] https://coderslegacy.com/python/delay-between-requests-in-scrapy/ [16] https://jrnold.github.io/r4ds-exercise-solutions/transform.html [17] https://levelup.gitconnected.com/implement-rate-limiting-in-python-d4f86b09259f [18] https://docs.python.org/3/faq/programming.html [19] https://www.javatpoint.com/how-to-add-time-delay-in-python [20] https://koji.mbox.centos.org/koji/buildinfo?buildID=22406 [21] https://pypi.org/project/ratelimit/ [22] https://docs.python.org/3/library/timeit.html [23] https://www.purplefrogsystems.com/2020/07/how-to-delay-a-python-loop/ [24] https://medium.com/clover-platform-blog/conquering-api-rate-limiting-dcac5552714d [25] https://learning.postman.com/docs/writing-scripts/pre-request-scripts/ [26] https://python-forum.io/thread-35631.html [27] https://365datascience.com/tutorials/python-tutorials/limit-rate-requests-web-scraping/ [28] https://instructobit.com/tutorial/108/How-to-share-global-variables-between-files-in-Python [29] https://learn.microsoft.com/en-us/azure/storage/blobs/storage-blobs-latency [30] https://dev.to/astagi/rate-limiting-using-python-and-redis-58gk [31] https://www.simplilearn.com/tutorials/python-tutorial/global-variable-in-python

  • ShareGPT @lemmy.fmhy.ml
    InternetPirate @lemmy.fmhy.ml

    limit python requests to 1.39 requests per second

    To limit your Python requests to 1.39 requests per second, you can use the time.sleep() function from the time module to introduce a delay between requests. You can also use third-party libraries like pyrate-limiter, requests-ratelimiter, or ratelimiter to achieve rate limiting.

    Here's an example using the time module:

     python
        
    import requests
    import time
    
    urls = [...]  # List of URLs to send requests to
    
    for url in urls:
        response = requests.get(url)
        print(response.status_code)
        time.sleep(1 / 1.39)  # Introduce a delay to achieve 1.39 requests per second
    
      

    Alternatively, you can use the pyrate-limiter library:

     python
        
    from pyrate_limiter import Duration, RequestRate, Limiter
    import requests
    
    rate_limit = RequestRate(1, Duration.SECOND / 1.39)  # 1 request per 1.39 seconds
    limiter = Limiter(rate_limit)
    
    urls = [...]  # List of URLs to send requests to
    
    for url in urls:
        with limiter:
            response = requests.get(url)
            print(response.status
      
    ShareGPT @lemmy.fmhy.ml
    InternetPirate @lemmy.fmhy.ml

    How to avoid defining sqlite3 conn and cursor in many functions across the code?

    To avoid defining the sqlite3.Connection and sqlite3.Cursor in many functions throughout your code, you can use the following approaches:

    1. Use a context manager: Create a context manager that handles the connection and cursor creation, as well as closing the connection when done. This way, you can use the with statement to manage the connection and cursor in your functions.
     python
        
    import sqlite3
    
    DB_FILE = "your_database_file.db"
    
    class DatabaseConnection:
        def __enter__(self):
            self.conn = sqlite3.connect(DB_FILE)
            self.cursor = self.conn.cursor()
            return self.cursor
    
        def __exit__(self, exc_type, exc_val, exc_tb):
            self.conn.commit()
            self.cursor.close()
            self.conn.close()
    
    def insert_post_to_db(issue: Issue, lemmy_post_id: int) -> None:
        with DatabaseConnection() as cursor:
            cursor.execute(
                "INSERT INTO posts (issue_url, lemmy_post_id, issue_title, issue_body) VALUES (?, ?, ?, ?)",
                (is
      
    Programming.dev Meta @programming.dev
    InternetPirate @lemmy.fmhy.ml

    Are AI-Generated Solutions Welcomed on programming.dev?

    I wanted to start a discussion about the use of AI-generated solutions on Programming.dev. Personally, I've found that AI-powered tools have been incredibly helpful in solving programming questions. I won't name any specific commercial software, but I use one that combines GPT-4 and web search to get more factual information. I write some answers I think I might revisit to the ShareGPT community, but I would prefer posting programming solutions to this instance. However, I'm not sure if AI-generated solutions are welcomed on programming.dev. I'd love to hear your thoughts on this. If AI-generated responses are accepted, how should we format the answers, should we just copy paste without quoting, should we quote the model, just mention that it's AI-generated,...?

  • Just change lemmy.post.create to lemmy.post.createe to trigger an AttributeError. That way you can debug the code without creating any posts. You can also use many print statements all around the code, I would use two for each line to make sure the computer isn't fooling you. Lastly, you can spin up your own Lemmy instance to not have to worry about the generated posts.

  • Lemmy Support @lemmy.ml
    InternetPirate @lemmy.fmhy.ml

    How to see the feed of another instance?

    I'm wondering if it's possible to see the local feed of another instance from the one I'm using. I'm interested in exploring content from other instances without having to visit every single community, but I'm not sure how to do it. I've tried searching for a way to do this on the documentation and using the Lemmy search, but I haven't found any clear instructions. Does anyone know how to see the local feed of another instance? Any help or guidance would be greatly appreciated!

    Lemmy Support @lemmy.ml
    InternetPirate @lemmy.fmhy.ml

    Does commenting on your own post bump it on the active filter view in Lemmy?

    In Lemmy, the active filter view is designed to prioritize posts with the latest activity, similar to how forums work. However, it remains unclear whether commenting on your own post in Lemmy will bump it on the active filter view. Some forum platforms, such as Discourse, allow a practice known as the "ghost bump," where users can make a post and delete it to draw attention to their post without adding new content[^1]. While it is uncertain if this is possible on Lemmy, it's worth noting that even if it were, it would result in an unnecessary comment that cannot be completely removed. The comment would still be visible, indicating that it was deleted by the post's creator. If you have any experience with Lemmy's active filter view or know whether commenting on your own post bumps it, please share your thoughts in the comments below.

    [^1]: What is "Bumping Topics"

    Lemmy @lemmy.ml
    InternetPirate @lemmy.fmhy.ml

    Seeking a Forgejo-hosted Mirror for Lemmy Outside the US

    As an enthusiastic supporter of Lemmy, I am eager to contribute to the project. However, I hold strong reservations about writing a single line of code for a project hosted on a Micro$oft server. While I have created a few issues on GitHub, I firmly believe that my contributions could be significantly amplified if there were a mirror of Lemmy that utilized Forgejo hosting outside the United States. I would be absolutely delighted to have the opportunity to contribute more actively to this incredible project if such an alternative hosting option were available.

    Singularity | Artificial Intelligence (ai), Technology & Futurology @lemmy.fmhy.ml
    InternetPirate @lemmy.fmhy.ml

    (Rumored Leak of) GPT-4 Architecture, Infrastructure, Training Dataset, Costs, Vision, MoE (11.07.2023)

    GPT-4's details are leaked.

    It is over.

    Everything is here: https://archive.is/2RQ8X

    Parameters count:

    GPT-4 is more than 10x the size of GPT-3. We believe it has a total of ~1.8 trillion parameters across 120 layers.

    Mixture Of Experts - Confirmed.

    OpenAI was able to keep costs reasonable by utilizing a mixture of experts (MoE) model. They utilizes 16 experts within their model, each is about ~111B parameters for MLP. 2 of these experts are routed to per forward pass.

    MoE Routing:

    While the literature talks a lot about advanced routing algorithms for choosing which experts to route each token to, OpenAI’s is allegedly quite simple, for the current GPT-4 model.

    There roughly ~55B shared parameters for attention.

    Inference:

    Each forward pass inference (generation of 1 token) only utilizes ~280B parameters and ~560 TFLOPs. This contrasts with the ~1.8 trillion parameters and ~3,700 TFLOP that would be required per forward pass of a purely dense model.

    **Datas

  • The paper actually demonstrates a 16-million context window with 92% accuracy. Most models can be retrained to have a 100k context window with over 92% accuracy, but the accuracy drops to 74% at 256k. The code has already been released on GitHub as well. I'm excited to see the development of 100k models using this method soon!

  • Singularity | Artificial Intelligence (ai), Technology & Futurology @lemmy.fmhy.ml
    InternetPirate @lemmy.fmhy.ml

    Focused Transformer: Contrastive Training for Context Scaling - 256k context length AI

    Summary:

    Focused Transformer: A new technique for long-context language modeling. The paper introduces Focused Transformer (FOT), a method that uses contrastive learning and external memory to improve the structure of the (key, value) space and extend the context length of transformer models. FOT can fine-tune existing large models without changing their architecture and achieve better performance on tasks that require long context.

    LONGLLAMA: Extending LLaMA’s context length with FOT. The paper demonstrates the application of FOT to fine-tune OpenLLaMA models, which are large language models with memory augmentation. The resulting models, called LONGLLAMAs, can handle a context length of up to 256k tokens and show improvements on few-shot learning tasks such as TREC and WebQS.

    Distraction issue: A key challenge for scaling context length. The paper identifies the distraction issue as a major obstacle for using large memory databases in multi-document scenarios. The distract

    No Stupid Questions @lemmy.world
    InternetPirate @lemmy.fmhy.ml

    Where can I seek medical advice online?

    Recently, I found myself questioning the accuracy of a diagnosis provided by a doctor I visited. Surprisingly, an AI seemed to offer a more insightful assessment. However, I understand the importance of not solely relying on AI-generated information. With that in mind, I'm eager to discover a reputable online platform where I can seek medical advice. Ideally, I hope to find a community where I can obtain multiple opinions to make a more informed decision about my health. If anyone could recommend such a site, I would greatly appreciate it.

    ShareGPT @lemmy.fmhy.ml
    InternetPirate @lemmy.fmhy.ml

    Price of insulin in China, Cuba, Laos, North Korea, and Vietnam?

    Information on the price of insulin in China, Cuba, Laos, North Korea, and Vietnam is not readily available in the search results. However, there is information on the cost of insulin in some other countries. According to the World Population Review, the cost of insulin varies widely by country, with the United States having the highest prices[1]. In 2018, RAND Corporation reported that the manufacturer price for any given type of insulin averaged five to ten times higher in the United States than in all other OECD countries[1]. Novo Nordisk, a major insulin manufacturer, has recently slashed the price of insulin in the United States[3].

    A study published in PMC in 2019 examined the prices, availability, and affordability of insulin in 13 low-income and middle-income countries. The study found that median procurement prices of human insulins across the countries ranged from $4.56 to $5.30, with large price variations seen across the countries[4]. The highest prices were found in two

    World News @lemmy.world
    InternetPirate @lemmy.fmhy.ml

    French Courts Are Giving Protesters INSANE Sentences

    French courts have been imposing disproportionately severe sentences for minor offenses, including 10 months in prison for stealing a can of Red Bull and one year for a homeless boy with schizophrenia caught looting a luxury store. The overwhelmed courts rush cases, provide minimal time for defendants, and prioritize punishment under the instruction of the Justice Minister. Furthermore, the French government is censoring social media and justifying it by claiming to protect public order, but it infringes upon free speech and mirrors tactics used by authoritarian regimes. The justice system exhibits a double standard, favoring the privileged, and creates a class divide, leading to unrest. Ironically, the government compares itself to oppressive nations while undermining democratic principles.

    World News @beehaw.org
    InternetPirate @lemmy.fmhy.ml

    French Courts Are Giving Protesters INSANE Sentences

    French courts have been imposing disproportionately severe sentences for minor offenses, including 10 months in prison for stealing a can of Red Bull and one year for a homeless boy with schizophrenia caught looting a luxury store. The overwhelmed courts rush cases, provide minimal time for defendants, and prioritize punishment under the instruction of the Justice Minister. Furthermore, the French government is censoring social media and justifying it by claiming to protect public order, but it infringes upon free speech and mirrors tactics used by authoritarian regimes. The justice system exhibits a double standard, favoring the privileged, and creates a class divide, leading to unrest. Ironically, the government compares itself to oppressive nations while undermining democratic principles.

    World News @lemmy.ml
    InternetPirate @lemmy.fmhy.ml

    French Courts Are Giving Protesters INSANE Sentences

    French courts have been imposing disproportionately severe sentences for minor offenses, including 10 months in prison for stealing a can of Red Bull and one year for a homeless boy with schizophrenia caught looting a luxury store. The overwhelmed courts rush cases, provide minimal time for defendants, and prioritize punishment under the instruction of the Justice Minister. Furthermore, the French government is censoring social media and justifying it by claiming to protect public order, but it infringes upon free speech and mirrors tactics used by authoritarian regimes. The justice system exhibits a double standard, favoring the privileged, and creates a class divide, leading to unrest. Ironically, the government compares itself to oppressive nations while undermining democratic principles.

    Singularity | Artificial Intelligence (ai), Technology & Futurology @lemmy.fmhy.ml
    InternetPirate @lemmy.fmhy.ml

    'Deepfake' Kanye Video Warns of Disinformation and Civil Unrest: 'AI Will Kill The Media Industry' (Yahoo article 05.07.2023)

    [YANDHI - WAR WITH THE MATRIX (KANYE AI X BIG BABY GANDHI)](https://youtube.com/watch?v=CGyPqImBOjY

    Singularity | Artificial Intelligence (ai), Technology & Futurology @lemmy.fmhy.ml
    InternetPirate @lemmy.fmhy.ml

    ChatGPT Code Interpreter: Image to Video Prompt (Midjourney) (video 09.07.2023)

  • I feel like this is what happened when you’d see posts with hundreds / thousands of upvotes but had only 20-ish comments.

    Nah it's the same here in Lemmy. It's because the algorithm only accounts for votes and not for user engagement.

  • Locked in a room with an internet connection? A lot. But without any contact with the outside world? Not nearly as much. It could have other people running experiments for it with an internet connection, but not without one.

    Anyway, whether or not the AGI can interact with the real world undermines the purpose of my explicit statement in the question. I specifically mentioned that it only operates as a human on a computer. I didn't mention it could acquire a physical body, so let's just assume it can't and can't use other people to do physical labor either.

  • I heard disruptive science is slowing down which I think means pretty much everything possible has already been thought of. So talking about things that exist, do you mean a cheaper solar panel or wind/water turbine? Or are we talking about science fiction like an Arc Reactor?