Llama 3.1 Megathread
Meta has released and open-sourced Llama 3.1 in three different sizes: 8B, 70B, and 405B
This new Llama iteration and update brings state-of-the-art performance to open-source ecosystems.
If you've had a chance to use Llama 3.1 in any of its variants - let us know how you like it and what you're using it for in the comments below!
Llama 3.1 Megathread
For this release, we evaluated performance on over 150 benchmark datasets that span a wide range of languages. In addition, we performed extensive human evaluations that compare Llama 3.1 with competing models in real-world scenarios. Our experimental evaluation suggests that our flagship model is competitive with leading foundation models across a range of tasks, including GPT-4, GPT-4o, and Claude 3.5 Sonnet. Additionally, our smaller models are competitive with closed and open models that have a similar number of parameters.

Image of the Week
HyperTech News Report #0002 - A New Challenger Approaches!
cross-posted from: https://lemmy.world/post/5965315
🤖 Happy FOSAI Friday! 🚀
Friday, September 29, 2023
HyperTech News Report #0002
Hello Everyone!
Welcome back to the HyperTech News Report! This week we're seeing some really exciting developments in futuristic technologies. With more tools and methods releasing by the day, I feel we're in for a renaissance in software. I hope hardware is soon to follow.. but I am here for it! So are you. Brace yourselves. Change is coming! This next year will be very interesting to watch unfold.
Table of Contents
Community Changelog
- Cleaned up some old content (let me know if you notice something that should be archived or updated)
Image of the Week
HyperTech News Report #0001 - Happy FOSAI Friday!
cross-posted from: https://lemmy.world/post/5549499
🤖 Happy FOSAI Friday! 🚀
Friday, September 22, 2023
HyperTech News Report #0001
Hello Everyone!
This series is a new vehicle for !fosai@lemmy.world news reports. In these posts I'll go over projects or news I stumble across week-over-week. I will try to keep Fridays consistent with this series, covering most of what I have been (but at regular cadence). For this week, I am going to do my best catching us up on a few old (and new) hot topics you may or may not have heard about already.
Table of Contents
Community Changelog
- Updated all resources on FOSAI ▲ XYZ.
- Added new content to FOSAI ▲ XYZ.
- Added new content and re
HyperTech News Report #0001 - Happy FOSAI Friday!
cross-posted from: https://lemmy.world/post/5549499
🤖 Happy FOSAI Friday! 🚀
Friday, September 22, 2023
HyperTech News Report #0001
Hello Everyone!
This series is a new vehicle for !fosai@lemmy.world news reports. In these posts I'll go over projects or news I stumble across week-over-week. I will try to keep Fridays consistent with this series, covering most of what I have been (but at regular cadence). For this week, I am going to do my best catching us up on a few old (and new) hot topics you may or may not have heard about already.
Table of Contents
Community Changelog
- Updated all resources on FOSAI ▲ XYZ.
- Added new content to FOSAI ▲ XYZ.
- Added new content and re
HyperTech News Report #0001 - Happy FOSAI Friday!
cross-posted from: https://lemmy.world/post/5549499
🤖 Happy FOSAI Friday! 🚀
Friday, September 22, 2023
HyperTech News Report #0001
Hello Everyone!
This series is a new vehicle for !fosai@lemmy.world news reports. In these posts I'll go over projects or news I stumble across week-over-week. I will try to keep Fridays consistent with this series, covering most of what I have been (but at regular cadence). For this week, I am going to do my best catching us up on a few old (and new) hot topics you may or may not have heard about already.
Table of Contents
Community Changelog
- Updated all resources on FOSAI ▲ XYZ.
- Added new content to FOSAI ▲ XYZ.
- Added new content and re
CodeLlama-34B - the First Open-Source Model Beating GPT-4 on HumanEvals
cross-posted from: https://lemmy.world/post/3879861
Beating GPT-4 on HumanEval with a Fine-Tuned CodeLlama-34B
Hello everyone! This post marks an exciting moment for !fosai@lemmy.world and everyone in the open-source large language model and AI community.
We appear to have a new contender on the block, a model apparently capable of surpassing OpenAI's state of the art ChatGPT-4 in coding evals (evaluations).
This is huge. Not too long ago I made an offhand comment on us catching up to GPT-4 within a year. I did not expect that prediction to end up being reality in half the time. Let's hope this isn't a one-off scenario and that we see a new wave of open-source models that begin to challenge OpenAI.
Buckle up, it's going to get interesting!
Here's some notes from the blog, which you should visit and read in its entirety:
Blog Post
Beating GPT-4 on HumanEval with a Fine-Tuned CodeLlama-34B
cross-posted from: https://lemmy.world/post/3879861
Beating GPT-4 on HumanEval with a Fine-Tuned CodeLlama-34B
Hello everyone! This post marks an exciting moment for !fosai@lemmy.world and everyone in the open-source large language model and AI community.
We appear to have a new contender on the block, a model apparently capable of surpassing OpenAI's state of the art ChatGPT-4 in coding evals (evaluations).
This is huge. Not too long ago I made an offhand comment on us catching up to GPT-4 within a year. I did not expect that prediction to end up being reality in half the time. Let's hope this isn't a one-off scenario and that we see a new wave of open-source models that begin to challenge OpenAI.
Buckle up, it's going to get interesting!
Here's some notes from the blog, which you should visit and read in its entirety:
Blog Post
Beating GPT-4 on HumanEval with a Fine-Tuned CodeLlama-34B
cross-posted from: https://lemmy.world/post/3879861
Beating GPT-4 on HumanEval with a Fine-Tuned CodeLlama-34B
Hello everyone! This post marks an exciting moment for !fosai@lemmy.world and everyone in the open-source large language model and AI community.
We appear to have a new contender on the block, a model apparently capable of surpassing OpenAI's state of the art ChatGPT-4 in coding evals (evaluations).
This is huge. Not too long ago I made an offhand comment on us catching up to GPT-4 within a year. I did not expect that prediction to end up being reality in half the time. Let's hope this isn't a one-off scenario and that we see a new wave of open-source models that begin to challenge OpenAI.
Buckle up, it's going to get interesting!
Here's some notes from the blog, which you should visit and read in its entirety:
Blog Post
Introducing Stable-Diffusion.cpp (Inference in Pure C/C++)
cross-posted from: https://lemmy.world/post/3549390
stable-diffusion.cpp
Introducing
stable-diffusion.cpp
, a pure C/C++ inference engine for Stable Diffusion! This is a really awesome implementation to help speed up home inference of diffusion models.Tailored for developers and AI enthusiasts, this repository offers a high-performance solution for creating and manipulating images using various quantization techniques and accelerated inference.
Key Features:
- Efficient Implementation: Utilizing plain C/C++, it operates seamlessly like llama.cpp and is built on the ggml framework.
- Multiple Precision Support: Choose between 16-bit, 32-bit float, and 4-bit to 8-bit integer quantization.
- Optimized Performance: Experience memory-efficient CPU inference with AVX, AVX2, and AVX512 support for x8
Cheetor - A New Multi-Modal LLM Strategy Empowered by Controllable Knowledge Re-Injection
cross-posted from: https://lemmy.world/post/3439370
Empowering Vision-Language Models to Follow Interleaved Vision-Language Instructions
A wild new GitHub Repo has appeared!
Today we cover Cheetah - an exciting new take on interleaving image and text context & instruction.
For higher quality images, please visit the main projects repo to see their code and approach in all of their glory.
I4 Benchmark
To facilitate research in interleaved vision-language instruction following, we build I4 (semantically Interconnected, Interleaved Image-Text Instruction-Following), an extensive large-scale benchmark of 31 tasks with diverse instructions in a unified instruction-response format, covering 20 diverse scenarios.
I4 has three important properties:
- Interleaved vision-language context: all the instructions contain sequenc
Cheetor - A New Multi-Modal LLM Strategy Empowered by Controllable Knowledge Re-Injection
cross-posted from: https://lemmy.world/post/3439370
Empowering Vision-Language Models to Follow Interleaved Vision-Language Instructions
A wild new GitHub Repo has appeared!
Today we cover Cheetah - an exciting new take on interleaving image and text context & instruction.
For higher quality images, please visit the main projects repo to see their code and approach in all of their glory.
I4 Benchmark
To facilitate research in interleaved vision-language instruction following, we build I4 (semantically Interconnected, Interleaved Image-Text Instruction-Following), an extensive large-scale benchmark of 31 tasks with diverse instructions in a unified instruction-response format, covering 20 diverse scenarios.
I4 has three important properties:
- Interleaved vision-language context: all the instructions contain sequenc
Incognito Pilot: The Next-Gen AI Code Interpreter for Sensitive Data
cross-posted from: https://lemmy.world/post/3350022
Incognito Pilot: The Next-Gen AI Code Interpreter for Sensitive Data
Hello everyone! Today marks the first day of a new series of posts featuring projects in my GitHub Stars.
Most of these repos are FOSS & FOSAI focused, meaning they should be hackable, free, and (mostly) open-source.
We're going to kick this series off by sharing Incognito Pilot. It’s like the ChatGPT Code Interpreter but for those who prioritize data privacy.
Project Summary from ChatGPT-4:
Features:
- Powered by Large Language Models like GPT-4 and Llama 2.
- Run code and execute tasks with Python interpreter.
- Privacy: Interacts with cloud but sensitive data stays local.
- Local or Remote: Choose between local LLMs
Incognito Pilot: The Next-Gen AI Code Interpreter for Sensitive Data
cross-posted from: https://lemmy.world/post/3350022
Incognito Pilot: The Next-Gen AI Code Interpreter for Sensitive Data
Hello everyone! Today marks the first day of a new series of posts featuring projects in my GitHub Stars.
Most of these repos are FOSS & FOSAI focused, meaning they should be hackable, free, and (mostly) open-source.
We're going to kick this series off by sharing Incognito Pilot. It’s like the ChatGPT Code Interpreter but for those who prioritize data privacy.
Project Summary from ChatGPT-4:
Features:
- Powered by Large Language Models like GPT-4 and Llama 2.
- Run code and execute tasks with Python interpreter.
- Privacy: Interacts with cloud but sensitive data stays local.
- Local or Remote: Choose between local LLMs
I used to feel the same way until I found some very interesting performance results from 3B and 7B parameter models.
Granted, it wasn’t anything I’d deploy to production - but using the smaller models to prototype quick ideas is great before having to rent a gpu and spend time working with the bigger models.
Give a few models a try! You might be pleasantly surprised. There’s plenty to choose from too. You will get wildly different results depending on your use case and prompting approach.
Let us know if you end up finding one you like! I think it is only a matter of time before we’re running 40B+ parameters at home (casually).
Vicuna v1.5 Has Been Released!
Click Here to be Taken to the Megathread!
from !fosai@lemmy.world
Vicuna v1.5 Has Been Released!
Shoutout to GissaMittJobb@lemmy.ml for catching this in an earlier post.
Given Vicuna was a widely appreciated member of the original Llama series, it'll be exciting to see this model evolve and adapt with fresh datasets and new training and fine-tuning approaches.
Feel free using this megathread to chat about Vicuna and any of your experiences with Vicuna v1.5!
Starting off with Vicuna v1.5
TheBloke is already sharing models!
Vicuna v1.5 GPTQ
7B
13B
**Vi
Vicuna v1.5 Has Been Released!
Click Here to be Taken to the Megathread!
from !fosai@lemmy.world
Vicuna v1.5 Has Been Released!
Shoutout to GissaMittJobb@lemmy.ml for catching this in an earlier post.
Given Vicuna was a widely appreciated member of the original Llama series, it'll be exciting to see this model evolve and adapt with fresh datasets and new training and fine-tuning approaches.
Feel free using this megathread to chat about Vicuna and any of your experiences with Vicuna v1.5!
Starting off with Vicuna v1.5
TheBloke is already sharing models!
Vicuna v1.5 GPTQ
7B
13B
**Vi
Vicuna v1.5 Has Been Released!
Click Here to be Taken to the Megathread!
from !fosai@lemmy.world
Vicuna v1.5 Has Been Released!
Shoutout to GissaMittJobb@lemmy.ml for catching this in an earlier post.
Given Vicuna was a widely appreciated member of the original Llama series, it'll be exciting to see this model evolve and adapt with fresh datasets and new training and fine-tuning approaches.
Feel free using this megathread to chat about Vicuna and any of your experiences with Vicuna v1.5!
Starting off with Vicuna v1.5
TheBloke is already sharing models!
Vicuna v1.5 GPTQ
7B
13B
**Vi
I am actively testing this out. It's hard to say at the moment. There's a lot to figure out deploying a model into a live environment, but I think there's real value in using them for technical tasks - especially as models mature and improve over time.
At the moment, though, performance is closer to GPT 3.5 than GPT 4, but I wouldn't be surprised if this is no longer the case within the next year or so.
After finally having a chance to test some of the new Llama-2 models, I think you're right. There's still some work to be done to get them tuned up... I'm going to dust off some of my notes and get a new index of those other popular gen-1 models out there later this week.
I'm very curious to try out some of these docker images, too. Thanks for sharing those! I'll check them when I can. I could also make a post about them if you feel like featuring some of your work. Just let me know!
Assuming everything from the papers translate into current platforms, yes! A rather significant one at that. Time will tell us the true results as people begin tinkering with this new approach in the near future.
Thanks for reading! I'm glad you enjoy the content. I find this tech beyond fascinating.
Who knows, over time you might even begin to pick up on some of the nuance you describe.
We're all learning this together!
Thanks for sharing this!
Good bot, I will do that next time.
Come hangout with us at !fosai@lemmy.world
I run this show solo at the moment, but do my best to keep everyone informed. I have much more content on the horizon. Would love to have you if we have what you're looking for.
FOSAI Posts:
OpenAI has launched a new initiative, Superalignment, aimed at guiding and controlling ultra-intelligent AI systems. Recognizing the imminent arrival of AI that surpasses human intellect, the project will dedicate significant resources to ensure these advanced systems act in accordance with human intent. It's a crucial step in managing the transformative and potentially dangerous impact of superintelligent AI.
I like to think this starts to explore interesting philosophical questions like human intent, consciousness, and the projection of will into systems that are far beyond our capabilities in raw processing power and input/output. What may happen from this intended alignment is yet to be seen, but I think we can all agree the last thing we want in these emerging intelligent machines is to do things we don't want them to do.
'Superalignment' is OpenAI's response in how to put up these safeguards. Whether or not this is the best method is to be determined.