Sebastian Raschka, PhD LinkedIn Posts

Learn how to post on LinkedIn like Sebastian Raschka, PhD. Learn from their content, engagement tactics, and network growth techniques

Sebastian Raschka, PhD's Best and Recent LinkedIn Posts
Hyperly.ai
profile picture of Sebastian Raschka, PhD
Sebastian Raschka, PhD
@sebastianraschka
about 1 year ago

One of the big bottlenecks with LLMs and Vision Transformers is GPU memory on consumer devices. I just wrote about my favorite techniques for

275 23 308
10 comments
profile picture of Sebastian Raschka, PhD
Sebastian Raschka, PhD
@sebastianraschka
7 months ago

Just wrote a new article on "Improving LoRA: Implementing Weight-Decomposed Low-Rank Adaptation (DoRA) from Scratch": https://lnkd.in/gJk48kms

I am super excited about DoRA, and based

495 29 544
20 comments
profile picture of Sebastian Raschka, PhD
Sebastian Raschka, PhD
@sebastianraschka
7 months ago

Just wrote a new article on "Improving LoRA: Implementing Weight-Decomposed Low-Rank Adaptation (DoRA) from Scratch": https://lnkd.in/gJk48kms

I am super excited about DoRA, and based

495 29 544
20 comments
profile picture of Sebastian Raschka, PhD
Sebastian Raschka, PhD
@sebastianraschka
7 months ago

While everyone is talking about Sora, there's a potential successor to LoRA (low-rank adaptation) called DoRA. Here's a closer look at the "DoRA:

1.3k 124 1.5k
44 comments
profile picture of Sebastian Raschka, PhD
Sebastian Raschka, PhD
@sebastianraschka
7 months ago

Another quality, well researched article by Sebastian Raschka, PhD.
Saved me a lot of time with finetuning various Huggingface LLMs for text classification

133 12 145
13 comments
profile picture of Sebastian Raschka, PhD
Sebastian Raschka, PhD
@sebastianraschka
7 months ago

Written & done! Just finished Chapter 4 on implementing an LLM architecture, which marks the marks the 50% point of the book.
(I

956 32 1.0k
42 comments
profile picture of Sebastian Raschka, PhD
Sebastian Raschka, PhD
@sebastianraschka
7 months ago

Can "small" finetuned LLMs with less than 2B parameters outperform larger openly available LLMs (Mixtral, Llama 2 Chat) and proprietary LLMs (ChatGPT)? Here's

622 50 713
41 comments
profile picture of Sebastian Raschka, PhD
Sebastian Raschka, PhD
@sebastianraschka
7 months ago

I've primarily focused on finetuning LLMs over the past few months and have recently started to explore model merging. As part of that

698 67 787
22 comments
profile picture of Sebastian Raschka, PhD
Sebastian Raschka, PhD
@sebastianraschka
8 months ago

In this episode of Leading With Data, we interact with Sebastian Raschka, PhD, AI Staff Educator at Lightning AI.

Sebastian Raschka, PhD is

71 8 80
1 comments
profile picture of Sebastian Raschka, PhD
Sebastian Raschka, PhD
@sebastianraschka
8 months ago

Looking at open source and research in 2024 so far, it seems we are moving towards making LLMs better (and smaller) without necessarily

833 85 936
18 comments
profile picture of Sebastian Raschka, PhD
Sebastian Raschka, PhD
@sebastianraschka
8 months ago

Proxy-tuning is a way to adapt LLMs without changing the model's weights. This is especially attractive if a given LLM is too resource-intensive

833 67 941
41 comments
profile picture of Sebastian Raschka, PhD
Sebastian Raschka, PhD
@sebastianraschka
8 months ago

There's a new promising method for finetuning LLMs without modifying their weights called proxy-tuning.
How does it work? It's a simple decoding-time method where you

2.3k 206 2.6k
87 comments
profile picture of Sebastian Raschka, PhD
Sebastian Raschka, PhD
@sebastianraschka
over 1 year ago

I just saw that my Ahead of AI magazine crossed the 20k subscriber mark!

I am incredibly grateful for all the support. Knowing

237 2 257
18 comments
profile picture of Sebastian Raschka, PhD
Sebastian Raschka, PhD
@sebastianraschka
over 1 year ago

Feeling a tad bit bored by decoder-only transformers like GPT and LLaMA? Let's delve back into the the world encoder-style BERT models for

307 12 340
21 comments
profile picture of Sebastian Raschka, PhD
Sebastian Raschka, PhD
@sebastianraschka
over 1 year ago

A new research paper just came out that proposes an alternative to reinforcement learning with human feedback (RLHF), which is used to finetune

701 68 788
19 comments
profile picture of Sebastian Raschka, PhD
Sebastian Raschka, PhD
@sebastianraschka
over 1 year ago

"LIMA: Less Is More for Alignment" might be a game-changer for researchers and tinkerers who want to develop capable LLMs.

In this paper,

623 56 699
20 comments
profile picture of Sebastian Raschka, PhD
Sebastian Raschka, PhD
@sebastianraschka
5 months ago

It's been many months in the making, and I am excited to share that the print version of Machine Learning Q and AI

1.4k 56 1.6k
73 comments
profile picture of Sebastian Raschka, PhD
Sebastian Raschka, PhD
@sebastianraschka
5 months ago

When doing machine learning and AI research (or writing books), making the code reproducible is usually desirable. Often, that's easier said than done!

433 19 480
28 comments
profile picture of Sebastian Raschka, PhD
Sebastian Raschka, PhD
@sebastianraschka
5 months ago

A good start to the new week: The highly requested Chapter 5 of my "Build an LLM from Scratch" book is finally available,

1.6k 101 1.8k
59 comments
profile picture of Sebastian Raschka, PhD
Sebastian Raschka, PhD
@sebastianraschka
5 months ago

We just added CodeGemma 7B support to LitGPT by a kind contribution from Andrei Aksionau.
Playing around with it for a bit, I think

353 22 394
19 comments
Hyperly
2024 - All rights reserved