
Customer Success Manager, Mid-Market
Still just vibes, still a BINGO.
Still just vibes, still a BINGO.
Since they're expecting candidates to operate on vibes only, so am I! And my vibes tell me this is a BINGO.
Can we just acknowledge that maybe it is recklessly, unforgivably irresponsible to allow 20-somethings who appear to have no professional experience before 2022 to be in charge of AI products that could destabilize democracies?
This is essentially a Director role with engineer duties, and right now, the salary is only compensating for one of those things.
Maybe it's just me, but it feels like they're asking this Director position to do a lot. It almost reads as if they asked ChatGPT for a Director of Support & Ops job description and then didn't whittle it down at all to fit this specific role.
This job seems fine if very, very corporate.
I continue to wish that Anthropic would address how they're mitigating the mental health risks of Trust & Safety work, and there are some minor flags ("fast-paced environment," hello darkness my old friend), but it's a solid Eh, It's Probably Fine.
Overall, I'm going to put this in Eh, It's Probably Fine, with a caveat that I really think anyone applying to this role ought to press hard on the "demonstrate grace under pressure" bit.
Can we not? Can we just not require a T&S professional to have a "fun" attitude, especially when you haven't given any space to explaining how you'll care for their mental well-being? Ugh. Honestly, that pisses me off enough to put this into Tread Carefully.
Genuinely one of the most diverse companies I've seen so far. And the rare case where a company claims diversity as a value and is clearly backing that up with their hiring. Cool!
Maybe they should have had Siena review this job listing before they posted it.
But I thought this is what Siena was for? Are you saying you need humans to support humans? Real food for thought.
The job description overall seems fine, but the pay is piddly for an onsite role in San Francisco, especially for a technical role. It's low enough, in fact, that I'm putting it in Tread Carefully.
Careers page is pretty basic; doesn't mention benefits at all, and neither does the job description. Otherwise, Cinder does a good job of explaining what its looking for in this role, and I don't see any major flags.
I was worried about doing this one, because I'm such a fan of the product, but Scribe's Careers page is really well done and the job description is mostly fine.
Given the listed duties and that this position reports to the VP of Operations and will collaborate closely with senior leadership, it really needs to be more senior than a Senior Manager (I'm thinking at least a Director of Support, if not Head of).
Salary range is wide but high enough that I don't think anyone will really care.
Y'all. Whenever a company mentions the actual benefits of a job as "in addition" to the PRIVILEGE and SPLENDOR of simply working for said company, as if being able to feed and provide for the health of your family is secondary to supporting an "iconic brand," well that is a major red flag.
I can't decide if quoting Albert Einstein in a job description is cute or weird. I'm leaning toward cute, because this actually seems like a really neat job, and I've detected no flags. Our first Green Means Go of this week!
It's a neat, actually useful product, but there are some definite red flags in the job description, so I advise caution and strategic questions if you end up interviewing.
Job description is thoughtful and well-written, benefits are excellent, and Careers page is clear and informative. This would be in Green Means Go except the salary range seems low for SF and NYC, especially considering they're wanting someone with a master's degree.
I always want to call attention to the fact that companies are very good at acknowledging when certain roles might come into contact with disturbing content but are very bad at addressing how they plan to support your mental health post-exposure to said content.
Same note as with the other T&S Lead role: companies are very good at acknowledging when certain roles might come into contact with disturbing content but are very bad at addressing how they plan to support your mental health post-exposure to said content.
Yes, it's AI. I'm as surprised as you.