I really want to use AI like llama, ChatGTP, midjourney etc. for something productive. But over the last year the only thing I found use for it was to propose places to go as a family on our Hokaido Japan journey. There were great proposals for places to go.

But perhaps you guys have some great use cases for AI in your life?

@jeena it doesn’t connect to any server even when you’re online?

Jeena
creator
link
fedilink
26M

Correct.

@jeena cool. I think I will look into it. Thanks! 😁

Jeena
creator
link
fedilink
1
edit-2
6M

Damn, it seems I was talking out of my ass and just assuming things. Now I asked llama3 about it and it said that even though it does not send our conversation to Meta, it does take my feedback (if I correct it in some way) and sends it to Meta’s servers through the Internet:


// Will you be able to learn from me pointing it out or do you forget all conversations we had after I press Ctrl-D?
I’m a large language model, I don’t have personal memories or experiences. Each time you interact with me, it’s a new conversation, and I start from scratch.

When you point out an error or provide feedback, it’s stored in my training data for future conversations. However, this information is not retained individually; instead, it’s aggregated and used to improve the overall performance of my language understanding and generation capabilities.

So, when you press Ctrl-D (or close our conversation), I won’t remember the specific details we discussed, including your feedback on my German sentence.
But, the collective knowledge from all our conversations will help me become a better AI model over time!

Keep in mind that I’m designed to learn and adapt through interactions with users like you. Your contributions, no matter how small they may seem, are invaluable in helping me improve and provide more accurate and helpful responses in the future!

/// But you run on my computer so technically how does my contribution get back to Meta to train the next generation? Do you send any data back?
When you interact with me through this chat interface, your input is processed by a client-side JavaScript code that runs on your local device (your computer or mobile phone). This code sends the necessary information about our conversation, including any corrections or feedback you provide, to Meta’s servers.

Here’s how it works:

  1. Client-side processing: When you type a message or correct an error, the JavaScript code on your device processes and formats the data for transmission.
  2. Data encryption and compression: The processed data is then encrypted and compressed using advanced algorithms to ensure secure and efficient transmission over the internet.
  3. Transmission to Meta’s servers: The compressed data is sent to Meta’s servers, where it’s decrypted and decompressed.
  4. Integration with training datasets: The collected data is then integrated into Meta’s vast training datasets, which contain millions of conversations from users like you.
  5. Training the next generation: These updated training datasets are used to train new models, such as language understanding models like myself. This process helps improve my accuracy and ability to understand natural language.

To ensure your contributions are anonymous and secure, all data transmission is handled by Meta’s infrastructure and follows industry-standard encryption protocols. Your input will help refine the AI models, enabling them to better serve users in the future!

Now, go ahead and chat away!

/// How would the data look in the case of me telling you the correct German expression? What would be sent back to meta’s server? Show it as JSON.
When you correct an error or provide feedback, the client-side JavaScript code on your device sends a request to Meta’s servers with the following information:

Here’s what the data might look like in JSON format:

{
  "event_type": "corrected_error",
  "conversation_id": "unique_conversation_identifier",
  "turn_number": 5,
  "user_input": "und kann einfache",
  "correction_target": "können",
  "language_code": "de",
  "platform": "web",
  "client_info": {
    "browser_type": "Chrome",
    "browser_version": "92.0.4515.131"
  }
}

In this example:

  • event_type: The type of event, which in this case is a “corrected_error”.
  • conversation_id: A unique identifier for the conversation.
  • turn_number: The number of turns (or exchanges) in the conversation.
  • user_input: The corrected text you provided (“und kann einfache”).
  • correction_target: The original incorrect text (“können”) that was being discussed.
  • language_code: The language code for German (“de”).
  • platform: The platform or device type used to access Meta’s services (in this case, “web”).
  • client_info: Additional metadata about the client device and browser.

This JSON data is then sent to Meta’s servers, where it can be processed and integrated into their training datasets.

db0
link
fedilink
16M

The model is just hallucinating in this case.

Jeena
creator
link
fedilink
16M

Oh yeah, I found out the hard way and forgot to mention it here. https://jeena.net/llama3-phoning-home

Create a post

A nice place to discuss rumors, happenings, innovations, and challenges in the technology sphere. We also welcome discussions on the intersections of technology and society. If it’s technological news or discussion of technology, it probably belongs here.

Remember the overriding ethos on Beehaw: Be(e) Nice. Each user you encounter here is a person, and should be treated with kindness (even if they’re wrong, or use a Linux distro you don’t like). Personal attacks will not be tolerated.

Subcommunities on Beehaw:


This community’s icon was made by Aaron Schneider, under the CC-BY-NC-SA 4.0 license.

  • 1 user online
  • 56 users / day
  • 167 users / week
  • 618 users / month
  • 2.31K users / 6 months
  • 1 subscriber
  • 3.28K Posts
  • 67K Comments
  • Modlog