Paper Note: ViT

ViT applies a standard Transformer directly to images ...

August 15, 2023 · 1 min · 183 words · Me

Paper Note: Attention is All You Need

The Transformer, the first sequence transduction model based entirely on attention, replacing the recurrent layers most commonly used in encoder-decoder architectures with multi-headed self-attention. ...

August 15, 2023 · 4 min · 699 words · Me

Paper Note: BERT

BERT is designed to pre-train deep bidirectional representations from unlabeled text by jointly conditioning on both left and right context in all layers. As a result, the pre-trained BERT model can be fine-tuned with just one additional output layer to create state-of-the-art models for a wide range of tasks. ...

August 15, 2023 · 4 min · 645 words · Me

A/B Testing Final Project

Implementation of Udacity A/B Testing course final project. ...

July 3, 2020 · 10 min · 2035 words · Me

Study Notes of Udacity A/B Testing

Study note of Udacity A/B Testing course. ...

July 3, 2020 · 35 min · 7355 words · Me