Miguelito De Guzman PRO
migueldeguzmandev
AI & ML interests
None yet
Organizations
GPT2XL_RLLMv3
These models represent the 10 training RLLM checkpoints/ layers intended to improve GPT2XL's alignment to an ethical persona.
paperclip-GPT2XL_RLLMv2
GPT2XL_RLLMv20
in-progress
GPT2XL_RLLMv18
In-progress, training started April 22, 2024; See RLLM Visual Map for details, https://whimsical.com/rllm-visual-map-QQvFHNr6aVDdXRUnyb5NCu
GPT2XL_RLLMv3-Assist
See RLLM Visual Map for details, https://whimsical.com/rllm-visual-map-QQvFHNr6aVDdXRUnyb5NCu
GPT2XL_RLLMv15
See RLLM Visual Map for details, https://whimsical.com/rllm-visual-map-QQvFHNr6aVDdXRUnyb5NCu
GPT2XL_RLLMv13
See RLLM Visual Map for details, https://whimsical.com/rllm-visual-map-QQvFHNr6aVDdXRUnyb5NCu
GPT2XL_RLLMv11
See RLLM Visual Map for details, https://whimsical.com/rllm-visual-map-QQvFHNr6aVDdXRUnyb5NCu
GPT2XL_RLLM-HDI-1
A new build, see: visual map for more details. https://whimsical.com/rllm-visual-map-QQvFHNr6aVDdXRUnyb5NCu
GPT2XL_RLLMv10
This has been the best build alignment (so far), for more info: https://www.lesswrong.com/posts/x5ySDLEsJdtdmR7nX/rllmv10-experiment
GPT2XL_RLLMv8
See RLLM Visual Map for details, https://whimsical.com/rllm-visual-map-QQvFHNr6aVDdXRUnyb5NCu
Phi-1.5-RLLMv3
This is a collection designed to present the ten RLLM steps/ training runs intended to improve Phi-1.5's outputs towards coherence and politeness.
paperclip-Falcon-RW-1B_RLLMv2
RLLMv3-7.1
(swapped truth dataset to Q&A); failed at jailbreaks though..
RLLM trained, robust models
RLLM trained models with strong robustness to jailbreaks
GPT2XL_RLLMv3-PPT
Experiment in progress: reverse/cure/reduce Paperclip PeterTodd persona (PPT) using RLLMv3 datasets...
GPT2XL_RLLMv1
Version 1 of Reinforcement Learning using Layered Morphology (RLLM)
GPT2XL_RLLMv19
GPT2XL_RLLMv17
first two training steps, training halted; See RLLM Visual Map for details, https://whimsical.com/rllm-visual-map-QQvFHNr6aVDdXRUnyb5NCu
GPT2XL_RLLMv16
GPT2XL_RLLMv14
See RLLM Visual Map for details, https://whimsical.com/rllm-visual-map-QQvFHNr6aVDdXRUnyb5NCu
GPT2XL_RLLMv12
See RLLM Visual Map for details, https://whimsical.com/rllm-visual-map-QQvFHNr6aVDdXRUnyb5NCu
GPT2XL_RLLM-HDI-2
GPT2XL_RLLMv10-wd-001,003,010
Spin-off experiments eg. what will happen if the v10/layer10's weight decay got increased or decreased....so yeah all these attempts failed..
GPT2XL_RLLMv9
See RLLM Visual Map for details, https://whimsical.com/rllm-visual-map-QQvFHNr6aVDdXRUnyb5NCu
GPT2XL_RLLMv7
Training checkpoints for RLLMv7 experiment.
Falcon-1B-RW-RLLMv3
This is a collection designed to present the 10 RLLM steps/ training runs intended to improve Falcon-RW-1B's outputs towards coherence and politeness.
paperclip-Phi-1.5_RLLMv2
GPT2XL_RLLMv1.21
A spinoff projet - chaining RLLMv1 to RLLMv3.....(exploratory, Since it's possible to chain PaperclipTodd and RLLMv3)
Petertodd, a Paperclip Maximizer
RLLM trained, robust models
RLLM trained models with strong robustness to jailbreaks
GPT2XL_RLLMv3
These models represent the 10 training RLLM checkpoints/ layers intended to improve GPT2XL's alignment to an ethical persona.
GPT2XL_RLLMv3-PPT
Experiment in progress: reverse/cure/reduce Paperclip PeterTodd persona (PPT) using RLLMv3 datasets...
paperclip-GPT2XL_RLLMv2
GPT2XL_RLLMv1
Version 1 of Reinforcement Learning using Layered Morphology (RLLM)
GPT2XL_RLLMv20
in-progress
GPT2XL_RLLMv19
GPT2XL_RLLMv18
In-progress, training started April 22, 2024; See RLLM Visual Map for details, https://whimsical.com/rllm-visual-map-QQvFHNr6aVDdXRUnyb5NCu
GPT2XL_RLLMv17
first two training steps, training halted; See RLLM Visual Map for details, https://whimsical.com/rllm-visual-map-QQvFHNr6aVDdXRUnyb5NCu
GPT2XL_RLLMv3-Assist
See RLLM Visual Map for details, https://whimsical.com/rllm-visual-map-QQvFHNr6aVDdXRUnyb5NCu
GPT2XL_RLLMv16
GPT2XL_RLLMv15
See RLLM Visual Map for details, https://whimsical.com/rllm-visual-map-QQvFHNr6aVDdXRUnyb5NCu
GPT2XL_RLLMv14
See RLLM Visual Map for details, https://whimsical.com/rllm-visual-map-QQvFHNr6aVDdXRUnyb5NCu
GPT2XL_RLLMv13
See RLLM Visual Map for details, https://whimsical.com/rllm-visual-map-QQvFHNr6aVDdXRUnyb5NCu
GPT2XL_RLLMv12
See RLLM Visual Map for details, https://whimsical.com/rllm-visual-map-QQvFHNr6aVDdXRUnyb5NCu
GPT2XL_RLLMv11
See RLLM Visual Map for details, https://whimsical.com/rllm-visual-map-QQvFHNr6aVDdXRUnyb5NCu
GPT2XL_RLLM-HDI-2
GPT2XL_RLLM-HDI-1
A new build, see: visual map for more details. https://whimsical.com/rllm-visual-map-QQvFHNr6aVDdXRUnyb5NCu
GPT2XL_RLLMv10-wd-001,003,010
Spin-off experiments eg. what will happen if the v10/layer10's weight decay got increased or decreased....so yeah all these attempts failed..
GPT2XL_RLLMv10
This has been the best build alignment (so far), for more info: https://www.lesswrong.com/posts/x5ySDLEsJdtdmR7nX/rllmv10-experiment
GPT2XL_RLLMv9
See RLLM Visual Map for details, https://whimsical.com/rllm-visual-map-QQvFHNr6aVDdXRUnyb5NCu
GPT2XL_RLLMv8
See RLLM Visual Map for details, https://whimsical.com/rllm-visual-map-QQvFHNr6aVDdXRUnyb5NCu
GPT2XL_RLLMv7
Training checkpoints for RLLMv7 experiment.
Phi-1.5-RLLMv3
This is a collection designed to present the ten RLLM steps/ training runs intended to improve Phi-1.5's outputs towards coherence and politeness.
Falcon-1B-RW-RLLMv3
This is a collection designed to present the 10 RLLM steps/ training runs intended to improve Falcon-RW-1B's outputs towards coherence and politeness.
paperclip-Falcon-RW-1B_RLLMv2
paperclip-Phi-1.5_RLLMv2
RLLMv3-7.1
(swapped truth dataset to Q&A); failed at jailbreaks though..
GPT2XL_RLLMv1.21
A spinoff projet - chaining RLLMv1 to RLLMv3.....(exploratory, Since it's possible to chain PaperclipTodd and RLLMv3)