본문 바로가기

전체 글

[Kaggle] 키 발급 https://velog.io/@skyepodium/Kaggle-API-%EC%82%AC%EC%9A%A9%EB%B2%95 Kaggle API 사용법Kaggle API 를 사용해서 1. 데이터 받기, 2. 제출을 해봅시다.velog.io 더보기
AlphaCodium https://github.com/Codium-ai/AlphaCodium GitHub - Codium-ai/AlphaCodium: Official implementation for the paper: "Code Generation with AlphaCodium: From Prompt EngineerinOfficial implementation for the paper: "Code Generation with AlphaCodium: From Prompt Engineering to Flow Engineering"" - Codium-ai/AlphaCodiumgithub.comAlphaCodium은 Codium AI에서 개발한 최첨단 코드 생성 도구입니다. 2022년 11월에 처음 공개되었으며, 인공지능(AI).. 더보기
[Kaggle] HMS - Harmful Brain Activity Classification 헝가리 출장을 갔다오는 바람에 제대로 참여하지 못한 ...https://www.kaggle.com/competitions/hms-harmful-brain-activity-classification/leaderboard?sort=published HMS - Harmful Brain Activity Classification | Kaggle www.kaggle.com모두가 방법은 비슷할 것 같았는데, 다 다른 방식인게 특이했다.내가 생각했던 방법과 가장 유사한 방법https://www.kaggle.com/competitions/hms-harmful-brain-activity-classification/discussion/492240 HMS - Harmful Brain Activity Classificatio.. 더보기
[안될공학 - IT 테크 신기술] 앞으로 AI 가 가져올 위험은 딥페이크보다 더 할 수 있습니다 (더밸류컨설팅 이병주 대표 4부) 좋은 사진만 나오면 결국 벡터가 확증편향되니까 앞으로 오히려 ai가 안좋아질 수 있다.좋은 데이터는 백터가 여러 방향으로 퍼지는게 좋은 데이터인공지능이 인공지능에 대한 결과를 학습을 하면 좋은 결과를 얻을 가능성 DOWN 더보기
트랜스포머 인코더 강의 https://www.youtube.com/watch?v=xrq2yN4K_-M 나쁘진 않지만, 초심자와 중심자 사이 난이도 같은 느낌? 더보기
Leave No Context Behind: Efficient Infinite Context Transformers with Infini attention https://www.youtube.com/watch?v=r_UBBfTPcF0https://arxiv.org/abs/2404.07143 Leave No Context Behind: Efficient Infinite Context Transformers with Infini-attentionThis work introduces an efficient method to scale Transformer-based Large Language Models (LLMs) to infinitely long inputs with bounded memory and computation. A key component in our proposed approach is a new attention technique dubbed.. 더보기
Apple Shocks Again: Introducing OpenELM Open Source AI Model That Changes Everything! https://www.youtube.com/watch?v=huH0fKmw0H0 Rmsnorm 사용, swiGLU 사용, Grouped query attention LR 공개 optimizer 공개 weight decay 공개 rmsnorm 사용 - 정확하지만 느림https://arxiv.org/abs/2404.14619 OpenELM: An Efficient Language Model Family with Open-source Training and Inference FrameworkThe reproducibility and transparency of large language models are crucial for advancing open research, ensuring the trustwort.. 더보기
Beyond A*: Better Planning with Transformers via Search Dynamics Bootstrapping (Searchformer) https://www.youtube.com/watch?v=PW4JiJ-WaY4https://arxiv.org/abs/2402.14083 Beyond A*: Better Planning with Transformers via Search Dynamics BootstrappingWhile Transformers have enabled tremendous progress in various application settings, such architectures still lag behind traditional symbolic planners for solving complex decision making tasks. In this work, we demonstrate how to train Transfor.. 더보기