🪴 My Second Brain

Search

SearchSearch
            • Beyond neural scaling laws beating power law scaling via data pruning
            • Deep Learning on a Data Diet Finding Important Examples Early in Training
            • SemDeDup Data-efficient learning at web-scale through semantic deduplication
              • LSH-MinHash
              • Memorization Without Overfitting Analyzing the Training Dynamics of Large Language Models
              • When Less is More Investigating Data Pruning for Pretraining LLMs at Scale
              • Language Models are Few-Shot Learners
              • The pile An 800gb dataset of diverse text for language modeling
            • D4 Improving LLM Pretraining via Document De-Duplication and Diversification
            • Scaling Laws for Neural Language Models
            • Textbooks are All You Need
          • Embedding
          • Inference
          • Language Models are Few-Shot Learners
          • Learning
          • Tokenization
          • 메모
          • 타임라인
            • Deep contextualized word representations
            • Explaining How Transformers Use Context to Build Predictions
            • Rethinking the Inception Architecture for Computer Vision
            • 작성 요령
      • Table
      • tools
      • Words
      • 사전 학습용 텍스트 데이터셋 평가
    Home

    ❯

    tools

    tools

    Mar 13, 2024, 1 min read

    https://.arxiv.org/abs/2303.09540 → https://ar5iv.org/abs/2303.09540 arxiv 문서를 html 형식으로 제공

    파파고 웹페이지 번역기 https://papago.naver.net/website

    Graph View

    Backlinks

    • No backlinks found

    Created with Quartz v4.2.2 © 2024

    • Github
    • Tistory