Home
Categories
EXPLORE
True Crime
Comedy
Business
Society & Culture
History
Sports
Health & Fitness
About Us
Contact Us
Copyright
ยฉ 2024 PodJoint
00:00 / 00:00
Sign in

or

Don't have an account?
Sign up
Forgot password
https://is1-ssl.mzstatic.com/image/thumb/Podcasts221/v4/52/ab/cb/52abcb67-3575-0960-7313-79789f23ad70/mza_547998439152404077.jpg/600x600bb.jpg
LlamaCast
Shahriar Shariati
49 episodes
4 months ago
Daily podcast about the published articles in the LLM field.
Show more...
Technology
News,
Tech News,
Science,
Mathematics
RSS
All content for LlamaCast is the property of Shahriar Shariati and is served directly from their servers with no modification, redirects, or rehosting. The podcast is not affiliated with or endorsed by Podjoint in any way.
Daily podcast about the published articles in the LLM field.
Show more...
Technology
News,
Tech News,
Science,
Mathematics
https://d3wo5wojvuv7l.cloudfront.net/t_rss_itunes_square_1400/images.spreaker.com/original/879177db874692a5aa0e7ad0353a362c.jpg
Qwen2.5-Coder
LlamaCast
24 minutes
12 months ago
Qwen2.5-Coder
๐Ÿ”ท Qwen2.5-Coder Technical Report

The report introduces the Qwen2.5-Coder series, which includes the Qwen2.5-Coder-1.5B and Qwen2.5-Coder-7B models. These models are specifically designed for coding tasks and have been pre-trained on a massive dataset of 5.5 trillion code-related tokens. A significant focus is placed on data quality, with detailed cleaning and filtering processes, and advanced training techniques such as file-level and repo-level pre-training. The models were rigorously tested on various benchmarks, including code generation, completion, reasoning, repair, and text-to-SQL tasks, where they demonstrated strong performance, even surpassing larger models in some areas. The report concludes with suggestions for future research, such as scaling model size and enhancing reasoning abilities.

๐Ÿ“Ž Link to paper

LlamaCast
Daily podcast about the published articles in the LLM field.