Skip to content

avera-codes/ds_resources_interviews

 
 

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

40 Commits
 
 

Repository files navigation

Data Science Resources

For interview preparation and learning

Table of Contents:

Interview Preparation

Questions

Data Science

Machine Learning

Deep Learning

SQL

NLP

Programming

Behavioural interview

Courses
Questions
Mock Interviews and Pieces of Advice

English

Home Assignments

Tips

Resources

Courses

Other

Algorithms and Data Structures

Platforms

Courses

Resources

Articles

Books

Python

Clean Code

Theory

Questions

Other

Practice

SQL

Courses

Practice

Machine Learning

Sites

Courses

Books

Cheetsheets

Articles

Applied ML

Blogs

Feature Engineering

Tutorials

Blog posts

Other

MLOps

General

Other

Deep Learning

Books

Courses

Tutorials

Blog posts

Other

Generative AI

NLP

Books

Courses

General

Large Language Models (LLMs) / Transformers

Reading papers with AI

Prompt Engineering

Tutorials

Blog posts

Articles

  • Word2Vec, Mikolov et al., Efficient Estimation of Word Representations in Vector Space
  • FastText, Bojanowski et al., Enriching Word Vectors with Subword Information
  • Attention, Bahdanau et al., Neural Machine Translation by Jointly Learning to Align and Translate
  • Transformers, Vaswani et al., Attention Is All You Need
  • BERT, Devlin et al., BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding
  • GPT-2, Radford et al., Language Models are Unsupervised Multitask Learners
  • GPT-3, Brown et al, Language Models are Few-Shot Learners
  • LaBSE, Feng et al., Language-agnostic BERT Sentence Embedding
  • CLIP, Radford et al., Learning Transferable Visual Models From Natural Language Supervision
  • RoPE, Su et al., RoFormer: Enhanced Transformer with Rotary Position Embedding
  • LoRA, Hu et al., LoRA: Low-Rank Adaptation of Large Language Models
  • InstructGPT, Ouyang et al., Training language models to follow instructions with human feedback
  • Scaling laws, Hoffmann et al., Training Compute-Optimal Large Language Models
  • FlashAttention, Dao et al., FlashAttention: Fast and Memory-Efficient Exact Attention with IO-Awareness
  • NLLB, NLLB team, No Language Left Behind: Scaling Human-Centered Machine Translation
  • Q8, Dettmers et al., LLM.int8(): 8-bit Matrix Multiplication for Transformers at Scale
  • Self-instruct, Wang et al., Self-Instruct: Aligning Language Models with Self-Generated Instructions
  • Alpaca, Taori et al., Alpaca: A Strong, Replicable Instruction-Following Model
  • LLaMA, Touvron, et al., LLaMA: Open and Efficient Foundation Language Models

Computer Vision

Graphs

Reinforcement Learning

RecSys

Courses

Books

Other

Time Series

Big Data

Books

Other

System Design

Machine Learning System Design

Math

General

Linear Algebra

Probability and Statistics

A/B Tests

General

Blog posts

Metrics

Other

Releases

No releases published

Packages

No packages published