I am actively looking for a tutor/advisor with expertise in AI x-risk, with the primary goal of collaboratively determining the most effective ways I can contribute to reducing AI existential risks (X-risk).
Tutoring Goals
I suspect that I misunderstand key components of the mental models that lead some highly rational and intelligent individuals to assign a greater than 50% probability of AI-related existential catastrophe (“p-doom”). By gaining a clearer understanding of these models, I aim to refine my thinking and make better-informed decisions about how to meaningfully reduce AI X-risk.
Specifically, I want to delve deeper into why and how misaligned AGI might be developed, and why it wouldn’t be straightforward to solve alignment before it becomes a critical issue.
To clarify, I do NOT believe we could contain or control a misaligned AGI with current safety practices. What I do find likely is that we will be able to avoid a situation altogether.
In addition to improving my understanding of AI X-risks, I also seek to explore strategies that I could aid in implementing in order to reduce AI X-risk.
About Me
- My primary motivation is effective altruism, and I believe that mitigating AI X-risk is the most important cause to work on. - I have 7 years of experience working with machine learning, with a focus on large language models (LLMs), and possess strong technical knowledge of the field. - My current p-doom estimate is 25%, derived from my own model, which gives about 5%, but I adjust upward in since some highly rational thinkers predicts significantly higher p-doom. Even if my p-doom were 1%, I would still view AI X-risk as the most pressing issue and dedicate my time to it.
Why Become My Tutor?
- You will be directly contributing to AI safety/alignment efforts, working with someone highly committed to making an impact. - Opportunity for **highly technical 1-on-1 discussions** about the cutting-edge in AI alignment and X-risk reduction strategies. - Compensation: $100–150 per hour (negotiable depending on your experience).
Ideal Qualifications
- Deep familiarity with AI existential risks and contemporary discussions surrounding AGI misalignment. - A genuine interest in refining mental models related to AI X-risk and collaborating on solutions. - p-doom estimate above 25%, since I aim to understand high p-doom perspectives. - Strong interpersonal compatibility: It’s crucial that we both find these discussions rewarding and intellectually stimulating.
Structure & Logistics
- Weekly one-hour meetings focused on deep discussions of AI X-risk, strategic interventions, and mental model refinement. - Flexible arrangement: you can invoice my company for the tutoring services.
How to Apply
If this opportunity sounds appealing to you, or if you know someone who may be a good fit, please DM me here on LessWrong.
[Question] Seeking AI Alignment Tutor/Advisor: $100–150/hr
I am actively looking for a tutor/advisor with expertise in AI x-risk, with the primary goal of collaboratively determining the most effective ways I can contribute to reducing AI existential risks (X-risk).
Tutoring Goals
I suspect that I misunderstand key components of the mental models that lead some highly rational and intelligent individuals to assign a greater than 50% probability of AI-related existential catastrophe (“p-doom”). By gaining a clearer understanding of these models, I aim to refine my thinking and make better-informed decisions about how to meaningfully reduce AI X-risk.
Specifically, I want to delve deeper into why and how misaligned AGI might be developed, and why it wouldn’t be straightforward to solve alignment before it becomes a critical issue.
To clarify, I do NOT believe we could contain or control a misaligned AGI with current safety practices. What I do find likely is that we will be able to avoid a situation altogether.
In addition to improving my understanding of AI X-risks, I also seek to explore strategies that I could aid in implementing in order to reduce AI X-risk.
About Me
- My primary motivation is effective altruism, and I believe that mitigating AI X-risk is the most important cause to work on.
- I have 7 years of experience working with machine learning, with a focus on large language models (LLMs), and possess strong technical knowledge of the field.
- My current p-doom estimate is 25%, derived from my own model, which gives about 5%, but I adjust upward in since some highly rational thinkers predicts significantly higher p-doom. Even if my p-doom were 1%, I would still view AI X-risk as the most pressing issue and dedicate my time to it.
Why Become My Tutor?
- You will be directly contributing to AI safety/alignment efforts, working with someone highly committed to making an impact.
- Opportunity for **highly technical 1-on-1 discussions** about the cutting-edge in AI alignment and X-risk reduction strategies.
- Compensation: $100–150 per hour (negotiable depending on your experience).
Ideal Qualifications
- Deep familiarity with AI existential risks and contemporary discussions surrounding AGI misalignment.
- A genuine interest in refining mental models related to AI X-risk and collaborating on solutions.
- p-doom estimate above 25%, since I aim to understand high p-doom perspectives.
- Strong interpersonal compatibility: It’s crucial that we both find these discussions rewarding and intellectually stimulating.
Structure & Logistics
- Weekly one-hour meetings focused on deep discussions of AI X-risk, strategic interventions, and mental model refinement.
- Flexible arrangement: you can invoice my company for the tutoring services.
How to Apply
If this opportunity sounds appealing to you, or if you know someone who may be a good fit, please DM me here on LessWrong.