#human Brewster
Explore tagged Tumblr posts
sakurastarkey · 4 months ago
Text
New Leaf cafe ☕️
3 notes · View notes
bearsandbaublesss · 6 months ago
Text
Tumblr media
Finally got around to finishing a piece of human Brewster from Animal Crossing that I sketched and lined? several months ago if not a year ago? lol
Now that I've finished this one, I kind of want to do another. Any suggestions on who I should draw as a human next?
3 notes · View notes
chaossmith2 · 9 months ago
Text
Tumblr media
Been replaying a lot of New Leaf so here are two of my favorite Special Characters I always enjoy interacting with
229 notes · View notes
blackbird-brewster · 10 months ago
Text
Tumblr media Tumblr media Tumblr media Tumblr media Tumblr media
Happy 55th Birthday, to the Gorgeous Human Peacock herself - - Paget V. Brewster
March 10, 1969
227 notes · View notes
retroautomaton · 2 years ago
Text
Tumblr media
☕️🌻🍮
504 notes · View notes
skillbattle · 2 years ago
Text
Tumblr media
barista pigeon man
489 notes · View notes
mollymarymarie · 2 days ago
Text
Series 12 will always be my favorite season of Bake Off. You mean to tell ME we were blessed with Chigs, Crystelle, Jurgen, Giuseppe, George, Rochica, Freya AND MOST IMPORTANTLY LIZZIE ALL IN ONE TENT???
The moment Lizzie said "I'm gonna stare him down" in her Scouse accent, I was immediately in love.
14 notes · View notes
number1spongebobfan · 2 months ago
Text
Tumblr media
Humanized Chuggington
@masterj
I think I'll call my au, "ChuggingKids" :)
13 notes · View notes
mikyapixie · 1 month ago
Text
4 years ago today, the Adventure Time Distant Lands episode The Obsidian premiered on HBO Max!!!
P2
I have the entire album of this episode on my Spotify playlist!!!🥰🥰🥰
13 notes · View notes
maxsarrt · 1 year ago
Text
Tumblr media
Is the animal crossing fandom even alive?
Isabelle and Brewster as humans !! (l'II probably do more characters in the future
Tumblr media
37 notes · View notes
oneluckcatsgirlhero-blog · 1 year ago
Text
Tumblr media
Original:
Tumblr media
7 notes · View notes
screechingnerdwombat · 2 years ago
Text
The Great Rivalry
I’ve been laid up in bed all week with covid and decided to read “Dracula” for the millionth? time  (hey who’s counting).  It’s a timeless story about good versus evil battling for dominance. Being the vampire geek that I am, I was again awed at the epic battle between Dracula and Van Helsing.  Dracula will stop at nothing to get what he wants from the humans around him, while Van Helsing will stop at nothing to protect them. This powerful human and vampire have become the paradigm for every human and vampire rivalry that has come after. 
One such rivalry that has always captivated me is Jerry Dandridge and Charley Brewster from “Fright Night.”  I admit that Jerry and Peter Vincent would be a better fit, but there’s always been a certain something about the struggle between Jerry and Charley that grips me in a such a way like no other. With the release of “Fright Night: Origins,” my interest in exploring their dynamic has piqued again.  Jerry and Charley represent the struggle between good versus evil, yes, but are so much more:  shadow and light, confidence and fear, corruption and innocence, vice and virtue, despair and hope, hate and love, respectively.  
 I could write a whole damn essay on this.  In fact, I think I will.
What are some of your favorite vampire/human rivalries in literature and on the big screen?  
3 notes · View notes
blackbird-brewster · 6 months ago
Photo
This fell out of my queue and now I just keep laughing in context of S17. Might I present Emily and Fuckin' Brian 🤣
Tumblr media Tumblr media Tumblr media Tumblr media
Frank and Sadie Doyle being absolutely perfect and adorable. (x)
433 notes · View notes
blackbird-brewster · 9 months ago
Text
Tumblr media
Paget "15% on the Kinsey Scale" Brewster
133 notes · View notes
mostlysignssomeportents · 2 months ago
Text
AI’s “human in the loop” isn’t
Tumblr media
I'll be in TUCSON, AZ from November 8-10: I'm the GUEST OF HONOR at the TUSCON SCIENCE FICTION CONVENTION.
Tumblr media
AI's ability to make – or assist with – important decisions is fraught: on the one hand, AI can often classify things very well, at a speed and scale that outstrips the ability of any reasonably resourced group of humans. On the other hand, AI is sometimes very wrong, in ways that can be terribly harmful.
Bureaucracies and the AI pitchmen who hope to sell them algorithms are very excited about the cost-savings they could realize if algorithms could be turned loose on thorny, labor-intensive processes. Some of these are relatively low-stakes and make for an easy call: Brewster Kahle recently told me about the Internet Archive's project to scan a ton of journals on microfiche they bought as a library discard. It's pretty easy to have a high-res scanner auto-detect the positions of each page on the fiche and to run the text through OCR, but a human would still need to go through all those pages, marking the first and last page of each journal and identifying the table of contents and indexing it to the scanned pages. This is something AI apparently does very well, and instead of scrolling through endless pages, the Archive's human operator now just checks whether the first/last/index pages the AI identified are the right ones. A project that could have taken years is being tackled with never-seen swiftness.
The operator checking those fiche indices is something AI people like to call a "human in the loop" – a human operator who assesses each judgment made by the AI and overrides it should the AI have made a mistake. "Humans in the loop" present a tantalizing solution to algorithmic misfires, bias, and unexpected errors, and so "we'll put a human in the loop" is the cure-all response to any objection to putting an imperfect AI in charge of a high-stakes application.
But it's not just AIs that are imperfect. Humans are wildly imperfect, and one thing they turn out to be very bad at is supervising AIs. In a 2022 paper for Computer Law & Security Review, the mathematician and public policy expert Ben Green investigates the empirical limits on human oversight of algorithms:
https://papers.ssrn.com/sol3/papers.cfm?abstract_id=3921216
Green situates public sector algorithms as the latest salvo in an age-old battle in public enforcement. Bureaucracies have two conflicting, irreconcilable imperatives: on the one hand, they want to be fair, and treat everyone the same. On the other hand, they want to exercise discretion, and take account of individual circumstances when administering justice. There's no way to do both of these things at the same time, obviously.
But algorithmic decision tools, overseen by humans, seem to hold out the possibility of doing the impossible and having both objective fairness and subjective discretion. Because it is grounded in computable mathematics, an algorithm is said to be "objective": given two equivalent reports of a parent who may be neglectful, the algorithm will make the same recommendation as to whether to take their children away. But because those recommendations are then reviewed by a human in the loop, there's a chance to take account of special circumstances that the algorithm missed. Finally, a cake that can be both had, and eaten!
For the paper, Green reviewed a long list of policies – local, national, and supra-national – for putting humans in the loop and found several common ways of mandating human oversight of AI.
First, policies specify that algorithms must have human oversight. Many jurisdictions set out long lists of decisions that must be reviewed by human beings, banning "fire and forget" systems that chug along in the background, blithely making consequential decisions without anyone ever reviewing them.
Second, policies specify that humans can exercise discretion when they override the AI. They aren't just there to catch instances in which the AI misinterprets a rule, but rather to apply human judgment to the rules' applications.
Next, policies require human oversight to be "meaningful" – to be more than a rubber stamp. For high-stakes decisions, a human has to do a thorough review of the AI's inputs and output before greenlighting it.
Finally, policies specify that humans can override the AI. This is key: we've all encountered instances in which "computer says no" and the hapless person operating the computer just shrugs their shoulders apologetically. Nothing I can do, sorry!
All of this sounds good, but unfortunately, it doesn't work. The question of how humans in the loop actually behave has been thoroughly studied, published in peer-reviewed, reputable journals, and replicated by other researchers. The measures for using humans to prevent algorithmic harms represent theories, and those theories are testable, and they have been tested, and they are wrong.
For example, people (including experts) are highly susceptible to "automation bias." They defer to automated systems, even when those systems produce outputs that conflict with their own expert experience and knowledge. A study of London cops found that they "overwhelmingly overestimated the credibility" of facial recognition and assessed its accuracy at 300% better than its actual performance.
Experts who are put in charge of overseeing an automated system get out of practice, because they no longer engage in the routine steps that lead up to the conclusion. Presented with conclusions, rather than problems to solve, experts lose the facility and familiarity with how all the factors that need to be weighed to produce a conclusion fit together. Far from being the easiest step of coming to a decision, reviewing the final step of that decision without doing the underlying work can be much harder to do reliably.
Worse: when algorithms are made "transparent" by presenting their chain of reasoning to expert reviewers, those reviewers become more deferential to the algorithm's conclusion, not less – after all, now the expert has to review not just one final conclusion, but several sub-conclusions.
Even worse: when humans do exercise discretion to override an algorithm, it's often to inject the very bias that the algorithm is there to prevent. Sure, the algorithm might give the same recommendation about two similar parents who are facing having their children taken away, but the judge who reviews the recommendations is more likely to override it for a white parent than for a Black one.
Humans in the loop experience "a diminished sense of control, responsibility, and moral agency." That means that they feel less able to override an algorithm – and they feel less morally culpable when they sit by and let the algorithm do its thing.
All of these effects are persistent even when people know about them, are trained to avoid them, and are given explicit instructions to do so. Remember, the whole reason to introduce AI is because of human imperfection. Designing an AI to correct human imperfection that only works when its human overseer is perfect produces predictably bad outcomes.
As Green writes, putting an AI in charge of a high-stakes decision, and using humans in the loop to prevent its harms, produces a "perverse effect": "alleviating scrutiny of government algorithms without actually addressing the underlying concerns." The human in the loop creates "a false sense of security" that sees algorithms deployed for high-stakes domains, and it shifts the responsibility for algorithmic failures to the human, creating what Dan Davies calls an "accountability sink":
https://profilebooks.com/work/the-unaccountability-machine/
The human in the loop is a false promise, a "salve that enables governments to obtain the benefits of algorithms without incurring the associated harms."
So why are we still talking about how AI is going to replace government and corporate bureaucracies, making decisions at machine speed, overseen by humans in the loop?
Well, what if the accountability sink is a feature and not a bug. What if governments, under enormous pressure to cut costs, figure out how to also cut corners, at the expense of people with very little social capital, and blame it all on human operators? The operators become, in the phrase of Madeleine Clare Elish, "moral crumple zones":
https://estsjournal.org/index.php/ests/article/view/260
As Green writes:
The emphasis on human oversight as a protective mechanism allows governments and vendors to have it both ways: they can promote an algorithm by proclaiming how its capabilities exceed those of humans, while simultaneously defending the algorithm and those responsible for it from scrutiny by pointing to the security (supposedly) provided by human oversight.
Tumblr media
Tor Books as just published two new, free LITTLE BROTHER stories: VIGILANT, about creepy surveillance in distance education; and SPILL, about oil pipelines and indigenous landback.
Tumblr media Tumblr media
If you'd like an essay-formatted version of this post to read or share, here's a link to it on pluralistic.net, my surveillance-free, ad-free, tracker-free blog:
https://pluralistic.net/2024/10/30/a-neck-in-a-noose/#is-also-a-human-in-the-loop
Tumblr media
Image: Cryteria (modified) https://commons.wikimedia.org/wiki/File:HAL9000.svg
CC BY 3.0 https://creativecommons.org/licenses/by/3.0/deed.en ==
289 notes · View notes
blackbird-brewster · 9 months ago
Text
Tumblr media Tumblr media Tumblr media Tumblr media Tumblr media Tumblr media
Criminal Minds on TikTok
73 notes · View notes