상세 컨텐츠

본문 제목

Machine Learning pipeline

Generative AI with Large Language Models

by Taeyoon.Kim.DS 2023. 3. 13. 20:23

본문

Hyperparameter Tuning --> Change hyperparameters can improve performance.

 

## Is this architectual optimal?

Why 512 neurons for dense? why dropout is 0.2? 

* Do the model need more or less hidden units to perfrom well?

* How does model size affect the convergence speed?

* Is there any trade off between converenge speed... etc

 

* Keras tuner built in search functionality.

16 --> 256 etc what is the best result?

 

Pip install -q -U keras -tuner

import kt

instead of 512, min_value = 16, max_value = 512 and step =16. Tuned by Keras tuner.

model.add(keras.layers.Dense(units=hp_units, activation='relu'))
Instead of 512. The optimized value is 64 so we can change it to 64. However, it might give you less accuracy.

 

 

define search strategy. 

kt.Hyperband strategy. search and baysian optimation and sklearn startegies.

Objective = 'val_accuracy'

 

stop_early = tf.keras.callbacks.EarlyStopping(monitor='val_loss', patience=5)

 

 

'Generative AI with Large Language Models' 카테고리의 다른 글

Generative AI & LLMs  (0) 2023.08.21
Introduction to LLMs and the generative AI project lifecycle  (0) 2023.08.21
AWS Innovate Data  (0) 2023.03.09
Advanced Testing Techniques  (0) 2023.03.07
What is Testing?  (0) 2023.03.07

관련글 더보기