Communities

Writing
Writing
Codidact Meta
Codidact Meta
The Great Outdoors
The Great Outdoors
Photography & Video
Photography & Video
Scientific Speculation
Scientific Speculation
Cooking
Cooking
Electrical Engineering
Electrical Engineering
Judaism
Judaism
Languages & Linguistics
Languages & Linguistics
Software Development
Software Development
Mathematics
Mathematics
Christianity
Christianity
Code Golf
Code Golf
Music
Music
Physics
Physics
Linux Systems
Linux Systems
Power Users
Power Users
Tabletop RPGs
Tabletop RPGs
Community Proposals
Community Proposals
tag:snake search within a tag
answers:0 unanswered questions
user:xxxx search by author id
score:0.5 posts with 0.5+ score
"snake oil" exact phrase
votes:4 posts with 4+ votes
created:<1w created < 1 week ago
post_type:xxxx type of post
Search help
Notifications
Mark all as read See all your notifications »
Q&A

Characterizing a sentient robot: sensory data

+0
−0

I have a sentient robot in my novel. Truth to be told, I have many. Sentience is somewhat cheap to achieve, meaning that there are multiple artificial beings that can be considered sentient by our standards.

I'm already making some differences and showing how he perceives the world through his set of sensory arrays. One of the core differences between us and a sentient machine, I imagine, would be sensory precision.

If I see a color, I may describe it picking between around 20-ish different terms. If I was trained all my life to distinguish between shades of colours, maybe I could get up to 60. But a sentient machine could - theoretically speaking - access raw data from its optic system. A robot could select exactly a range of pixels from his optical "nerve" and return an hexadecimal value that represent the shade with far more precision.

"Bring me the faint yellow dress, please."

"Oh, you mean the #EEFEEF one?"

(Worse still if the robot is encoding in some finer format, like rgb: "Wait, I see only a rgb(255,255,250) dress here!")

Coming to my question: I was thinking of characterizing my robotic PoV using this heightened sensory data. Example of this could be him commenting on the exact weight of an object he lifts, the exact distance between his location and a point he has to reach, and so on.

Is this a good idea, or would it be tiring for the reader?

History
Why does this post require attention from curators or moderators?
You might want to add some details to your flag.
Why should this post be closed?

0 comment threads

2 answers

You are accessing this answer with a direct link, so it's being shown above all other answers regardless of its score. You can return to the normal view.

+0
−0

I suggest this long answer of mine (90 votes) on a similar topic; it will define some of the terms you are using.

A "sentient" or self-aware being (machine or biological) will have an internal model of itself in the world, and be able to model (or simulate) with relative success how its own actions will cause changes in the world.

Most sentient beings would realize that using RGB codes with a human is pointless, their internal model of humans (necessary for them to work properly) will know that humans don't distinguish colors, weights, distances, etc with any precision. So it is not realistic for them to use these when talking with humans.

Edit: I should point out; "self-awareness" does not imply "emotional". We have self-aware robots already; self-driving cars and other robots that navigate a natural environment, or that need to be careful not to bump into or hit people or things. It only implies an ability to represent itself as the one object it can control in an environment of other (fixed or moving) objects it cannot: It is aware of itself.

Also, as a professional artist once informed me, nothing is ever one color. Even on a clear summer day, the sky is not "blue", it is fifteen shades of blue, and my shirt is not red, it is at least five shades of red depending on lighting, shadows, folds and wrinkles. So even the robot would know that the yellow dress is NOT #EEFEEF, but is a whole spectrum of colors, and predominately or on average what a human would term "yellow." In order to be effective, that would be built into its AI, it wouldn't constantly say things the humans cannot understand and be confused by their confusion.

That said, robot to robot, they might be precise and say "walk five thousand, nine hundred and eighty three feet."

Also, if a human requests it, they might report the exact value of their sensors, for sound-level, air pressure, temperature, humidity, distance (or distance walked), weight, altitude, compass direction, etc. Just like I might check my GPS coordinates on my iPhone.

Personally I think this kind of commentary would be tiring for the reader. It might be fun a few times for a curious human to ask this sort of thing, I can only imagine a 2-3 year old's endless "why" questions posed to an encyclopedic and willing care-giver robot. Of course the robot is not an endless fount, either, eventually like a human it would have to admit it doesn't know how it knows something, or would have to cite a resource it believes is true. But the chain would be longer, and it might be entertaining if you pick the right starting question.

History
Why does this post require attention from curators or moderators?
You might want to add some details to your flag.

0 comment threads

+0
−0

As @Amadeus points out, a robot programmed to interact with humans would know what range of colours "yellow" corresponds to, and would use "yellow" when interacting with humans. Interacting with other robots, a robot might find it more comfortable to use the specific wavelength, or some similar representation. I can easily imagine an AI being more comfortable with precise information than with an approximation.

However, there is a third option: your sentient robot might wish to be obnoxious. In that case, insisting on this precision, showing off their superiority compared to humans, would be fitting.

Would it tire the reader? Only if you over-use it. You might remember, in the original Star Trek, Spock had a gimmick - he was overly precise with calculations, and any mathematical figures. It showed up no more than once or twice per episode, not in every episode. More than that, and it would have been too much. This sprinkling of precision was just enough to maintain Spock firmly in the "stranger" slot; his precision was non-human.

Your robots' precision, just as Spock's precision, might come in useful. Once you've established it, a character might make use of it on occasion, asking a robot for a precise figure.

History
Why does this post require attention from curators or moderators?
You might want to add some details to your flag.

0 comment threads

Sign up to answer this question »