🐱 Skyspace3.0

Search

SearchSearch
      • a note on effective altruism et al
      • acting NOW
      • actual explanation
      • addictive technologies are harmful
      • bids
      • Can I enjoy arbitrary things?
      • change is atomic
      • deep learning class -- some things I've learned
      • dissonance. fun. conviction.
      • dont look up
      • double standard
      • Emotional Health in the face of the AI Situation
      • good arguments
      • on feelings
      • Preventing AI's from Taking Over the World
      • Real Hashing
      • responsibility reprise
      • stakes
      • stimuli
      • taboo
      • taking life seriously
      • taking social initiative
      • talking to people
      • the world is full of allies
      • The Wrong Person
      • Thinking about people's reactions to AI xrisk
      • worrying about giving offense
      • alignment drift notes
      • alignment-drift-empirics
      • backdoors take 2
      • concrete questions
      • Does deliberative alignment incentivize lying?
      • How scary is HHH?
      • LPE Agenda
      • MAD Agenda
      • Mechanistic Probability Estimation
      • project ideas 04.21.25
      • questions 02.16.25
      • questions 02.20.25
      • questions 03.10.25
      • questions 3.22
      • Quick Thoughts on Ensembling Detectors
      • refuting NCP
      • safe distillation
      • safe distillation FAILED
      • Solving AI Safety from 1st principles
      • BPP in Sigma2
      • Computer Security Notes
      • cryptographic proofs
      • cuckoo hashing
      • DQN Maze Solver
      • EM reprise
      • Goldreich-Levin Theorem
      • Hirahara Shimizu Planted Clique Paper Summary
      • interesting ideas
      • IQPs + friends
      • iterated amplification
      • Kolmogorov complexity
      • low vs high stakes alignment
      • murphy -- Theoretical ML textbook
      • online convex optimization
      • P2P communication
      • particle filter
      • predictive AI
      • RL
      • Ryan
      • smoothing backdoors
        • 2024-06-21
        • 2024-06-25
        • 2024-06-27
        • 2024-07-24
        • Borůvka
        • Expansion Review
        • hashtable without hash functions
        • network decompositions
        • routing
      • A letter to the Senator
      • A mistake -- too critical of Anthropic specifically
      • A simple alignment proposal
      • ai discussion slideshow
      • algorithm class notes
      • alignment meta-research algorithm
      • arena notes
      • backdoors and deceptive alignment
      • Brainstorming Some Open Questions
      • bucket list of learning exercises
      • CAIP model organism proposal
      • Causal Scrubbing Notes
      • Check if Alek released a new post script
      • crypto primitives
      • defaults
      • Do you have the Markov property?
      • Ethics of Computing Essay 1
      • finding c4s
      • fine tuning
      • good blog posts
      • inference
      • intro-to-rl
      • Inverse Reward Design -- paper summary
      • mildly hot takes
      • mnist
      • off-switch
      • OLD --- ARC formal verification notes
      • open-ness
      • physics notes
      • qmhw9
      • raising awareness
      • ridiculous BFS
      • Risks from AI -- elevator pitch
      • scalable alignment
      • school tidbits
      • some ml projects and reflection
      • some permutation avoidance notes
      • some problems from my friends
      • some random super old alignment notes
      • superposition
      • talking to myself
      • the end, again
      • The Inner Alignment Problem
      • two mistakes in my plans for mitigating xrisk
      • What could go wrong?
      • What Should We Do About the xRisk Posed by ASI?
      • wins
      • yearbookv2
      • Addressing some more common counterarguments
      • Lemma 1
      • Lemma 2
      • Lemma 3
      • Lemma 4
      • main
      • normative claims
      • 5 years
      • 100% is easier than 99%
      • accept the truth
      • AI xrisk
      • akrasia
      • avoid_discomfort.cpp
      • be specific
      • Being intellectually curious
      • Can I change my objective function? Should I?
      • cognitive distortion list
      • commitments
      • consciousness
      • contact
      • cooking
      • determinism is useless
      • Doing good work
      • existence
      • existence of objective function
      • Favorite Endeavors by Aesthetics
      • flake
      • food-water
      • fun activity ideas
      • good problems
      • good subgoals
      • Goodhart's law
      • goodness(universe)
      • how I want to spend my life
      • human compute is scarce
      • intention
      • is judging people a good idea?
      • justification
      • laughing at serious things
      • listening
      • live research options
      • mind reading
      • motivated reasoning
      • not fail with abandon
      • nukes xrisk thoughts
      • OLD AI xrisk argument
      • on writing, blogging and being opinionated
      • outer loop, inner loop
      • pascal-wager
      • point of research
      • prediction vs explanation
      • press the button
      • procrastinating choices
      • questions
      • recovering from being tired when working
      • responsibility
      • science AI xrisk article
      • self betrayal
      • sleep
      • social
      • some random projects
      • speak the truth?
      • Terms for others in my utility function
      • trivia
      • walking
      • Want to X or want to have X'd
      • What do I need money for?
      • what influence do perceived expectations have on me?
      • Will humans beat death in the next 30 years?
      • win-the-lottery-utility
      • working healthily
      • wrong exists
      • your work is not your worth
    Home

    ❯

    0stack

    Folder: 0stack

    27 items under this folder.

    • Sep 29, 2025

      Real Hashing

      • Jul 27, 2025

        stakes

        • Jul 17, 2025

          Preventing AI's from Taking Over the World

          • Jun 29, 2025

            dissonance. fun. conviction.

            • Jun 29, 2025

              on feelings

              • Jun 29, 2025

                The Wrong Person

                • Jun 29, 2025

                  addictive technologies are harmful

                  • Jun 29, 2025

                    taking social initiative

                    • life
                  • Jun 29, 2025

                    talking to people

                    • life
                  • Jun 29, 2025

                    Can I enjoy arbitrary things?

                    • Jun 29, 2025

                      Thinking about people's reactions to AI xrisk

                      • life
                    • Jun 29, 2025

                      a note on effective altruism et al

                      • Jun 29, 2025

                        acting NOW

                        • Jun 29, 2025

                          actual explanation

                          • Jun 29, 2025

                            bids

                            • Jun 29, 2025

                              change is atomic

                              • life
                            • Jun 29, 2025

                              good arguments

                              • Jun 29, 2025

                                responsibility reprise

                                • Jun 29, 2025

                                  stimuli

                                  • todo
                                • Jun 29, 2025

                                  taboo

                                  • philosophy
                                • Jun 29, 2025

                                  the world is full of allies

                                  • Jun 29, 2025

                                    worrying about giving offense

                                    • Jan 26, 2025

                                      Emotional Health in the face of the AI Situation

                                      • life
                                    • Jan 17, 2025

                                      dont look up

                                      • Dec 21, 2024

                                        double standard

                                        • Nov 09, 2024

                                          deep learning class -- some things I've learned

                                          • Sep 19, 2024

                                            taking life seriously


                                            Created with Quartz v4.2.3 © 2025

                                            • GitHub
                                            • Discord Community