I think of it two ways: a) Perhaps machines/other complicated systems (nebulas, centers of stars, Internet) already HAVE consciousness and awareness. b) If they do not and if there is some sort of tipping point of complexity required, i. What is that tipping point? ii. How will we know if it has consciousness? iii. How will we know if it has awareness? iv. How will we know if it has self-awareness? I'm skeptical of claims of AI-as-conscious-awareness/awareness-of-self based upon complexity alone, because if that is the case, then it's likely that MANY things are already conscious. I would also have no problem with the concept of crystals being intelligent and having consciousness for example. But I don't think we can legitimately argue complexity-alone as 'somehow' bringing forth automatic awareness / consciousness / self-awareness *if* we start with the assumption that: a) only humans have it now (which is quite ridiculous, considering animals) b) Only humans and some animals have it, but do not below a certain point. (more reasonable, but some of the tests are rather ridiculous, like the mirror test). Personally? I believe all life forms do have consciousness; a social consciousness from which the individual may or may not be formed. Bacteria, for example, have a social awareness and, by extension, _must_ have some sort of individual awareness for they respond to the group, whether the group is "like us" or "not like us" based upon chemical signally. But.. you can't have a "me" without first an "us". That's my thinking.