99国产精品欲av蜜臀,可以直接免费观看的AV网站,gogogo高清免费完整版,啊灬啊灬啊灬免费毛片

網易首頁 > 網易號 > 正文 申請入駐

AI 工程師轉型路線圖!資源匯總

0
分享至


大家好,我是i學習的老章

周末了,推薦一個新項目——AI工程師轉型路線圖

tips:搭配之前我推薦的幾個工具一起食用,效果更佳


項目地址:https://github.com/InterviewReady/ai-engineering-resources
Tokenization 分詞處理
  • Byte-pair Encoding
    https://arxiv.org/pdf/1508.07909

  • Byte Latent Transformer: Patches Scale Better Than Tokens
    https://arxiv.org/pdf/2412.09871

Vectorization 向量化處理
  • BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding
    https://arxiv.org/pdf/1810.04805

  • IMAGEBIND: One Embedding Space To Bind Them All
    https://arxiv.org/pdf/2305.05665

  • SONAR: Sentence-Level Multimodal and Language-Agnostic Representations
    https://arxiv.org/pdf/2308.11466

  • FAISS library
    https://arxiv.org/pdf/2401.08281

  • Facebook Large Concept Models
    https://arxiv.org/pdf/2412.08821v2

Infrastructure 基礎設施
  • TensorFlow
    https://arxiv.org/pdf/1605.08695

  • Deepseek filesystem
    https://github.com/deepseek-ai/3FS/blob/main/docs/design_notes.md

  • Milvus DB
    https://www.cs.purdue.edu/homes/csjgwang/pubs/SIGMOD21_Milvus.pdf

  • Billion Scale Similarity Search : FAISS
    https://arxiv.org/pdf/1702.08734

  • Ray
    https://arxiv.org/abs/1712.05889

Core Architecture 核心架構
  • Attention is All You Need
    https://papers.neurips.cc/paper/7181-attention-is-all-you-need.pdf

  • FlashAttention
    https://arxiv.org/pdf/2205.14135

  • Multi Query Attention
    https://arxiv.org/pdf/1911.02150

  • Grouped Query Attention
    https://arxiv.org/pdf/2305.13245

  • Google Titans outperform Transformers
    https://arxiv.org/pdf/2501.00663

  • VideoRoPE: Rotary Position Embedding
    https://arxiv.org/pdf/2502.05173

Mixture of Experts 專家混合模型
  • Sparsely-Gated Mixture-of-Experts Layer
    https://arxiv.org/pdf/1701.06538

  • GShard
    https://arxiv.org/abs/2006.16668

  • Switch Transformers
    https://arxiv.org/abs/2101.03961

RLHF 基于人類反饋的強化學習
  • Deep Reinforcement Learning with Human Feedback
    https://arxiv.org/pdf/1706.03741

  • Fine-Tuning Language Models with RHLF
    https://arxiv.org/pdf/1909.08593

  • Training language models with RHLF
    https://arxiv.org/pdf/2203.02155

Chain of Thought 思維鏈
  • Chain-of-Thought Prompting Elicits Reasoning in Large Language Models
    https://arxiv.org/pdf/2201.11903

  • Chain of thought
    https://arxiv.org/pdf/2411.14405v1/

  • Demystifying Long Chain-of-Thought Reasoning in LLMs
    https://arxiv.org/pdf/2502.03373

Reasoning 推理
  • Transformer Reasoning Capabilities
    https://arxiv.org/pdf/2405.18512

  • Large Language Monkeys: Scaling Inference Compute with Repeated Sampling
    https://arxiv.org/pdf/2407.21787

  • Scale model test times is better than scaling parameters
    https://arxiv.org/pdf/2408.03314

  • Training Large Language Models to Reason in a Continuous Latent Space
    https://arxiv.org/pdf/2412.06769

  • DeepSeek R1
    https://arxiv.org/pdf/2501.12948v1

  • A Probabilistic Inference Approach to Inference-Time Scaling of LLMs using Particle-Based Monte Carlo Methods
    https://arxiv.org/pdf/2502.01618

  • Latent Reasoning: A Recurrent Depth Approach
    https://arxiv.org/pdf/2502.05171

  • Syntactic and Semantic Control of Large Language Models via Sequential Monte Carlo
    https://arxiv.org/pdf/2504.13139

Optimizations 優化方案
  • The Era of 1-bit LLMs: All Large Language Models are in 1.58 Bits
    https://arxiv.org/pdf/2402.17764

  • FlashAttention-3: Fast and Accurate Attention with Asynchrony and Low-precision
    https://arxiv.org/pdf/2407.08608

  • ByteDance 1.58
    https://arxiv.org/pdf/2412.18653v1

  • Transformer Square
    https://arxiv.org/pdf/2501.06252

  • Inference-Time Scaling for Diffusion Models beyond Scaling Denoising Steps
    https://arxiv.org/pdf/2501.09732

  • 1b outperforms 405b
    https://arxiv.org/pdf/2502.06703

  • Speculative Decoding
    https://arxiv.org/pdf/2211.17192

Distillation 蒸餾
  • Distilling the Knowledge in a Neural Network
    https://arxiv.org/pdf/1503.02531

  • BYOL - Distilled Architecture
    https://arxiv.org/pdf/2006.07733

  • DINO
    https://arxiv.org/pdf/2104.14294

SSMs 狀態空間模型
  • RWKV: Reinventing RNNs for the Transformer Era
    https://arxiv.org/pdf/2305.13048

  • Mamba
    https://arxiv.org/pdf/2312.00752

  • Transformers are SSMs: Generalized Models and Efficient Algorithms Through Structured State Space Duality
    https://arxiv.org/pdf/2405.21060

  • Distilling Transformers to SSMs
    https://arxiv.org/pdf/2408.10189

  • LoLCATs: On Low-Rank Linearizing of Large Language Models
    https://arxiv.org/pdf/2410.10254

  • Think Slow, Fast
    https://arxiv.org/pdf/2502.20339

Competition Models 競賽模型
  • Google Math Olympiad 2
    https://arxiv.org/pdf/2502.03544

  • Competitive Programming with Large Reasoning Models
    https://arxiv.org/pdf/2502.06807

  • Google Math Olympiad 1
    https://www.nature.com/articles/s41586-023-06747-5

Hype Makers
  • Can AI be made to think critically
    https://arxiv.org/pdf/2501.04682

  • Evolving Deeper LLM Thinking
    https://arxiv.org/pdf/2501.09891

  • LLMs Can Easily Learn to Reason from Demonstrations Structure
    https://arxiv.org/pdf/2502.07374

Hype Breakers
  • Separating communication from intelligence
    https://arxiv.org/pdf/2301.06627

  • Language is not intelligence
    https://gwern.net/doc/psychology/linguistics/2024-fedorenko.pdf

Image Transformers 圖像轉換器
  • Image is 16x16 word
    https://arxiv.org/pdf/2010.11929

  • CLIP
    https://arxiv.org/pdf/2103.00020

  • deepseek image generation
    https://arxiv.org/pdf/2501.17811

Video Transformers 視頻轉換器
  • ViViT: A Video Vision Transformer
    https://arxiv.org/pdf/2103.15691

  • Joint Embedding abstractions with self-supervised video masks
    https://arxiv.org/pdf/2404.08471

  • Facebook VideoJAM ai gen
    https://arxiv.org/pdf/2502.02492

Case Studies 案例分析
  • Automated Unit Test Improvement using Large Language Models at Meta
    https://arxiv.org/pdf/2402.09171

  • Retrieval-Augmented Generation with Knowledge Graphs for Customer Service Question Answering
    https://arxiv.org/pdf/2404.17723v1

  • OpenAI o1 System Card
    https://arxiv.org/pdf/2412.16720

  • LLM-powered bug catchers
    https://arxiv.org/pdf/2501.12862

  • Chain-of-Retrieval Augmented Generation
    https://arxiv.org/pdf/2501.14342

  • Swiggy Search
    https://bytes.swiggy.com/improving-search-relevance-in-hyperlocal-food-delivery-using-small-language-models-ecda2acc24e6

  • Swarm by OpenAI
    https://github.com/openai/swarm

  • Netflix Foundation Models
    https://netflixtechblog.com/foundation-model-for-personalized-recommendation-1a0bd8e02d39

  • Model Context Protocol
    https://www.anthropic.com/news/model-context-protocol

  • uber queryGPT
    https://www.uber.com/en-IN/blog/query-gpt/


最后推薦一個最近我在學習的大模型課程


特別聲明:以上內容(如有圖片或視頻亦包括在內)為自媒體平臺“網易號”用戶上傳并發布,本平臺僅提供信息存儲服務。

Notice: The content above (including the pictures and videos if any) is uploaded and posted by a user of NetEase Hao, which is a social media platform and only provides information storage services.

相關推薦
熱點推薦
你上下行速度多少!工信部:1000Mbps及以上接入速率的固網寬帶接入達2.23億戶

你上下行速度多少!工信部:1000Mbps及以上接入速率的固網寬帶接入達2.23億戶

快科技
2025-06-25 16:18:11
63歲吳鎮宇新造型太棒了,沒有老人味還越來越年輕,至少年輕20歲

63歲吳鎮宇新造型太棒了,沒有老人味還越來越年輕,至少年輕20歲

TVB的四小花
2025-06-16 02:57:47
杜蘭特攜手威少加盟火箭,掘金甩約基奇,火箭重組沖冠局

杜蘭特攜手威少加盟火箭,掘金甩約基奇,火箭重組沖冠局

秘密清風
2025-06-27 14:01:20
李湘和16歲女兒回國,直奔商場買鞋買衣服,王詩齡不化妝很清純

李湘和16歲女兒回國,直奔商場買鞋買衣服,王詩齡不化妝很清純

檸檬有娛樂
2025-06-27 10:13:46
追求標準答案危害無窮

追求標準答案危害無窮

水寒說語文
2025-06-26 17:30:16
黃一鳴自曝:王思聰每次約她,車費都給10萬,來給5萬,回再給5萬

黃一鳴自曝:王思聰每次約她,車費都給10萬,來給5萬,回再給5萬

漢史趣聞
2025-06-24 10:07:59
楊瀚森自評現役模板:同為16順位申京 吃煎餅調侃和周琦夢幻聯動

楊瀚森自評現役模板:同為16順位申京 吃煎餅調侃和周琦夢幻聯動

醉臥浮生
2025-06-27 15:48:22
財政部:截至5月末全國地方政府債務余額512475億元

財政部:截至5月末全國地方政府債務余額512475億元

澎湃新聞
2025-06-27 10:36:06
艱難抉擇!湖北一高二考生被屏蔽,走還是不走?家長和學校犯難了

艱難抉擇!湖北一高二考生被屏蔽,走還是不走?家長和學校犯難了

火山詩話
2025-06-26 06:42:34
當年為什么查辦褚時健?

當年為什么查辦褚時健?

顧史
2025-06-09 20:26:22
又一個經濟奇跡——孟加拉國

又一個經濟奇跡——孟加拉國

古老板的老巢
2025-06-24 23:41:45
博伊爾斯女友社媒曬照慶祝男友被猛龍選中:哦嘿呀

博伊爾斯女友社媒曬照慶祝男友被猛龍選中:哦嘿呀

懂球帝
2025-06-27 08:40:19
洪森惱羞成怒,竟曝光英拉秘事,佩通坦絕地反擊,泰專機直飛中國

洪森惱羞成怒,竟曝光英拉秘事,佩通坦絕地反擊,泰專機直飛中國

嘆知
2025-06-23 10:34:33
中國游戲超越日本!來自日本玩家の認可!IGN日本高度評價《歸唐》!甘肅文旅廳宣布支持開發組古跡掃描

中國游戲超越日本!來自日本玩家の認可!IGN日本高度評價《歸唐》!甘肅文旅廳宣布支持開發組古跡掃描

Steam社區
2025-06-26 11:57:21
絕了!碘伏才是隱藏的省錢刺客!我居然白白燒了五年冤枉錢!

絕了!碘伏才是隱藏的省錢刺客!我居然白白燒了五年冤枉錢!

墻頭草
2025-04-01 12:00:11
昔日女神成方圓老了!兩段婚姻失敗無子女,對愛失望選擇單身

昔日女神成方圓老了!兩段婚姻失敗無子女,對愛失望選擇單身

一盅情懷
2025-06-27 15:54:59
1小時賣了578億!小米YU7大定28.9萬輛遠超SU7,小米ADR暴漲12%,雷軍連喊三個“安全”

1小時賣了578億!小米YU7大定28.9萬輛遠超SU7,小米ADR暴漲12%,雷軍連喊三個“安全”

金融界
2025-06-26 23:50:34
塵埃落定,重返巴薩?內馬爾官宣,1年合同,年薪1500萬

塵埃落定,重返巴薩?內馬爾官宣,1年合同,年薪1500萬

東球弟
2025-06-27 12:34:08
笑暈!江蘇衛視又逮著常州一個勁嚯嚯!這回“遭殃”的你卻想不到

笑暈!江蘇衛視又逮著常州一個勁嚯嚯!這回“遭殃”的你卻想不到

深析古今
2025-06-27 17:14:39
萬萬沒想到!拒給馬筱梅做婚紗,緬懷大S的蘭玉,竟是他的前女友

萬萬沒想到!拒給馬筱梅做婚紗,緬懷大S的蘭玉,竟是他的前女友

史行途
2025-06-27 15:07:11
2025-06-27 21:44:49
機器學習與Python社區 incentive-icons
機器學習與Python社區
機器學習算法與Python
3014文章數 11024關注度
往期回顧 全部

科技要聞

雷軍:小米汽車成功沒靠營銷,靠的是能力

頭條要聞

美國打擊伊朗核設施后 美軍一將領進入特朗普核心圈

頭條要聞

美國打擊伊朗核設施后 美軍一將領進入特朗普核心圈

體育要聞

曼城“庫里”連線,送尤文晚安好夢

娛樂要聞

炸裂!榜一大姐深夜怒錘頂流

財經要聞

合新鐵路建設材料以次充好 多家單位被罰

汽車要聞

配置升級/貴賓座椅 全新GL8陸上公務艙售22.99萬

態度原創

手機
時尚
旅游
本地
軍事航空

手機要聞

小米MIX Flip2線下上手:體驗后,說說真實感受

到了夏天,中年女人穿過膝裙不配運動鞋,才能美得優雅又得體

旅游要聞

熱聞|清明假期將至,熱門目的地有哪些?

本地新聞

被貴妃帶火的“唐代頂流”,如今怎么不火了

軍事要聞

美媒揭美軍37小時奔襲伊朗細節

無障礙瀏覽 進入關懷版 主站蜘蛛池模板: 石景山区| 丰顺县| 陆川县| 微山县| 重庆市| 威海市| 阆中市| 马关县| 三门县| 信阳市| 当雄县| 华宁县| 隆昌县| 商都县| 突泉县| 卓尼县| 深水埗区| 龙里县| 金塔县| 都江堰市| 万全县| 鄂温| 湄潭县| 岑巩县| 开封市| 荥阳市| 高清| 商南县| 平阴县| 辛集市| 安多县| 漳州市| 阿拉尔市| 堆龙德庆县| 琼中| 屏南县| 灵寿县| 虞城县| 华亭县| 平舆县| 元谋县|