Cognitive scientists

597 quotes found

"I am not shooting at immortality through my books, no. Nor do I think Chopin was shooting at immortality through his music. That strikes me as a very selfish goal, and I don't think Chopin was particularly selfish. I would also say that I think that music comes much closer to capturing the essence of a composer's soul than do a writer's ideas capture the writer's soul. Perhaps some very emotional ideas that I express in my books can get across a bit of the essence of my soul to some readers, but I think that Chopin's music probably does a lot better job (and the same holds, of course, for many composers). I personally don't have any thoughts about "shooting for immortality" when I write. I try to write simply in order to get ideas out there that I believe in and find fascinating, because I'd like to let other people be able share those ideas. But intellectual ideas alone, no matter how fascinating they are, are not enough to transmit a soul across brains. Perhaps, as I say, my autobiographical passages — at least some of them — get tiny shards of my soul across to some people. But such autobiographical story-telling is not nearly as effective a means of soul-transmission as is living with someone you love for many years of your lives, and sharing profound life goals with them — that's for sure!"

- Douglas Hofstadter

0 likesAcademics from the United StatesCognitive scientistsMathematicians from the United StatesPhilosophers from the United StatesPoets from the United States
"The distinction between responsible moral agents and beings with diminished or no responsibility is coherent, real, and important. It is coherent, even if in many instances it is hard to apply; it draws an empirically real line, in that we don't all fall on one side; and, most important, the distinction matters: the use we make of it plays a crucial role in the quality and meaning of our lives. [...] We want to hold ourselves and others responsible, but we recognize that our intuitions often support the judgement that a particular individual has "diminished responsibility" because of his or her infirmities, or because of particularly dire circumstances upon upbringing or at the time of action. We also find it plausible to judge that nonhuman animals, infants, and those who are severely handicapped mentally are not responsible at all. But since we are all more or less imperfect, will there be anyone left to be responsible after we have excused all those with good excuses? [...] We must set up some efficiently determinable threshold for legal competence, never for a moment supposing that there couldn't be intuitively persuasive "counterexamples" to whatever line we draw, but declaring in advance that such pleas will not be entertained. [...] The effect of such an institution [...] is to create [...] a class of legally culpable agents whose subsequent liability to punishment maintains the credibility of the sanctions of the laws. The institution, if it is to maintain itself, must provide for the fine tuning of its arbitrary thresholds as new information (or misinformation) emerges that might undercut its credibility. One can speculate that there is an optimal setting of the competence threshold (for any particular combination of social circumstances, degree of public sophistication, and so on) that maximizes the bracing effect of the law. A higher than optimal threshold would encourage a sort of malingering on the part of the defendants, which, if recognized by the populace, would diminish their respect for the law and hence diminish its deterrent effect. And a lower than optimal threshold would yield a diminishing return of deterrence and lead to the punishment of individuals who, in the eyes of society, "really couldn't help it." The public perception of the fairness of the law is a critical factor in its effectiveness."

- Daniel Dennett

0 likesAcademics from the United StatesPhilosophers from the United StatesCognitive scientistsNon-fiction authors from the United StatesBiographers from the United States
"A neurosurgeon once told me about operating on the brain of a young man with epilepsy. As is customary in this kind of operation, the patient was wide awake, under only local anesthesia, while the surgeon delicately explored his exposed cortex, making sure that the parts tentatively to be removed were not absolutely vital by stimulating them electrically and asking the patient what he experienced. Some stimulations provoked visual flashes or hand-raisings, others a sort of buzzing sensation, but one spot produced a delighted response from the patient: "It's 'Outta Get Me' by Guns N' Roses, my favorite heavy metal [sic] band!"I asked the neurosurgeon if he had asked the patient to sing or hum along with the music, since it would be fascinating to learn how "high fidelity" the provoked memory was. Would it be in exactly the same key and tempo as the record? Such a song (unlike "Silent Night") has one canonical version, so we could simply have superimposed a recording of the patient's humming with the standard record and compare the results. Unfortunately, even though a tape recorder had been running during the operation, the surgeon hadn't asked the patient to sing along. "Why not?" I asked, and he replied: "I hate rock music!"Later in the conversation the neurosurgeon happened to remark that he was going to have to operate again on the same young man, and I expressed the hope that he would just check to see if he could restimulate the rock music, and this time ask the fellow to sing along. "I can't do that," replied the neurosurgeon, "since I cut out that part." "It was part of the epileptic focus?" I asked, and he replied, "No, I already told you – I hate rock music.""

- Daniel Dennett

0 likesAcademics from the United StatesPhilosophers from the United StatesCognitive scientistsNon-fiction authors from the United StatesBiographers from the United States
"Minds are in limited supply, and each mind has a limited capacity for memes, and hence there is considerable competition among memes for entry in as many minds as possible. This competition is the major selective force in the memosphere, and, just as in the biosphere, the challenge has been met with great ingenuity. For instance, whatever virtues (from our perspective) the following memes have, they have in common the property of having phenotypic expressions that tend to make their own replication more likely by disabling or preempting the environmental forces that would tend to extinguish them: the meme for faith, which discourages the exercise of the sort of critical judgment that might decide that the idea of faith was, all things considered a dangerous idea; the meme for tolerance or free speech; the meme of including in a chain letter a warning about the terrible fates of those who have broken the chain in the past; the conspiracy theory meme, which has a built-in response to the objection that there is no good evidence of a conspiracy: "Of course not – that's how powerful the conspiracy is!" Some of these memes are "good" perhaps and others "bad"; what they have in common is a phenotypic effect that systematically tends to disable the selective forces arrayed against them. Other things being equal, population memetics predicts that conspiracy theory memes will persist quite independently of their truth, and the meme for faith is apt to secure its own survival, and that of the religious memes that ride piggyback on it, in even the most rationalistic environments. Indeed, the meme for faith exhibits frequency-dependent fitness: it flourishes best when it is outnumbered by rationalistic memes; in an environment with few skeptics, the meme for faith tends to fade from disuse."

- Daniel Dennett

0 likesAcademics from the United StatesPhilosophers from the United StatesCognitive scientistsNon-fiction authors from the United StatesBiographers from the United States
"In a Thumbnail Sketch here is [the Multiple Drafts theory of consciousness] so far:There is no single, definitive "stream of consciousness," because there is no central Headquarters, no Cartesian Theatre where "it all comes together" for the perusal of a Central Meaner. Instead of such a single stream (however wide), there are multiple channels in which specialist circuits try, in parallel pandemoniums, to do their various things, creating Multiple Drafts as they go. Most of these fragmentary drafts of "narrative" play short-lived roles in the modulation of current activity but some get promoted to further functional roles, in swift succession, by the activity of a virtual machine in the brain. The seriality of this machine (its "von Neumannesque" character) is not a "hard-wired" design feature, but rather the upshot of a succession of coalitions of these specialists.The basic specialists are part of our animal heritage. They were not developed to perform peculiarly human actions, such as reading and writing, but ducking, predator-avoiding, face-recognizing, grasping, throwing, berry-picking, and other essential tasks. They are often opportunistically enlisted in new roles, for which their talents may more or less suit them. The result is not bedlam only because the trends that are imposed on all this activity are themselves part of the design. Some of this design is innate, and is shared with other animals. But it is augmented, and sometimes even overwhelmed in importance, by microhabits of thought that are developed in the individual, partly idiosyncratic results of self-exploration and partly the predesigned gifts of culture. Thousands of memes, mostly borne by language, but also by wordless "images" and other data structures, take up residence in an individual brain, shaping its tendencies and thereby turning it into a mind."

- Daniel Dennett

0 likesAcademics from the United StatesPhilosophers from the United StatesCognitive scientistsNon-fiction authors from the United StatesBiographers from the United States
"I have grown accustomed to the disrespect expressed by some of the participants for their colleagues in the other disciplines. "Why, Dan," ask the people in artificial intelligence, "do you waste your time conferring with those neuroscientists? They wave their hands about 'information processing' and worry about where it happens, and which neurotransmitters are involved, but they haven't a clue about the computational requirements of higher cognitive functions." "Why," ask the neuroscientists, "do you waste your time on the fantasies of artificial intelligence? They just invent whatever machinery they want, and say unpardonably ignorant things about the brain." The cognitive psychologists, meanwhile, are accused of concocting models with neither biological plausibility nor proven computational powers; the anthropologists wouldn't know a model if they saw one, and the philosophers, as we all know, just take in each other's laundry, warning about confusions they themselves have created, in an arena bereft of both data and empirically testable theories. With so many idiots working on the problem, no wonder consciousness is still a mystery. All these charges are true, and more besides, but I have yet to encounter any idiots. Mostly the theorists I have drawn from strike me as very smart people – even brilliant people, with the arrogance and impatience that often comes with brilliance – but with limited perspectives and agendas, trying to make progress on the hard problems by taking whatever shortcuts they can see, while deploring other people's shortcuts. No one can keep all the problems and details clear, including me, and everyone has to mumble, guess and handwave about large parts of the problem."

- Daniel Dennett

0 likesAcademics from the United StatesPhilosophers from the United StatesCognitive scientistsNon-fiction authors from the United StatesBiographers from the United States
"[I]f you want to reason about faith, and offer a reasoned (and reason-responsive) defense of faith as an extra category of belief worthy of special consideration, I'm eager to [participate]. I certainly grant the existence of the phenomenon of faith; what I want to see is a reasoned ground for taking faith as a way of getting to the truth, and not, say, just as a way people comfort themselves and each other (a worthy function that I do take seriously). But you must not expect me to go along with your defense of faith as a path to truth if at any point you appeal to the very dispensation you are supposedly trying to justify. Before you appeal to faith when reason has you backed into a corner, think about whether you really want to abandon reason when reason is on your side. You are sightseeing with a loved one in a foreign land, and your loved one is brutally murdered in front of your eyes. At the trial it turns out that in this land friends of the accused may be called as witnesses for the defense, testifying about their faith in his innocence. You watch the parade of his moist-eyed friends, obviously sincere, proudly proclaiming their undying faith in the innocence of the man you saw commit the terrible deed. The judge listens intently and respectfully, obviously more moved by this outpouring than by all the evidence presented by the prosecution. Is this not a nightmare? Would you be willing to live in such a land? Or would you be willing to be operated on by a surgeon you tells you that whenever a little voice in him tells him to disregard his medical training, he listens to the little voice? I know it passes in polite company to let people have it both ways, and under most circumstances I wholeheartedly cooperate with this benign agreement. But we're seriously trying to get at the truth here, and if you think that this common but unspoken understanding about faith is anything better than socially useful obfuscation to avoid mutual embarrassment and loss of face, you have either seen much more deeply into the issue that any philosopher ever has (for none has ever come up with a good defense of this) or you are kidding yourself."

- Daniel Dennett

0 likesAcademics from the United StatesPhilosophers from the United StatesCognitive scientistsNon-fiction authors from the United StatesBiographers from the United States
"A faith, like a species, must evolve or go extinct when the environment changes. It is not a gentle process in either case. … It's nice to have grizzly bears and wolves living in the wild. They are no longer a menace; we can peacefully co-exist, with a little wisdom. The same policy can be discerned in our political tolerance, in religious freedom. You are free to preserve or create any religious creed you wish, so long as it does not become a public menace. We're all on the Earth together, and we have to learn some accommodation. … The message is clear: those who will not accommodate, who will not temper, who insist on keeping only the purest and wildest strain of their heritage alive, we will be obliged, reluctantly, to cage or disarm, and we will do our best to disable the memes they fight for. Slavery is beyond the pale. Child abuse is beyond the pale. Discrimination is beyond the pale. The pronouncing of death sentences on those who blaspheme against a religion (complete with bounties or reward for those who carry them out) is beyond the pale. It is not civilized, and it is owed no more respect in the name of religious freedom than any other incitement to cold-blooded murder. … That is — or, rather, ought to be, the message of multiculturalism, not the patronizing and subtly racist hypertolerance that "respects" vicious and ignorant doctrines when they are propounded by officials of non-European states and religions."

- Daniel Dennett

0 likesAcademics from the United StatesPhilosophers from the United StatesCognitive scientistsNon-fiction authors from the United StatesBiographers from the United States
"Evolution embodies information in every part of every organism. … This information doesn't have to be copied into the brain at all. It doesn't have to be "represented" in "data structures" in the nervous system. It can be exploited by the nervous system, however, which is designed to rely on, or exploit, the information in the hormonal systems just as it is designed to rely on, or exploit, the information embodied in your limbs and eyes. So there is wisdom, particularly about preferences, embodied in the rest of the body. By using the old bodily systems as a sort of sounding board, or reactive audience, or critic, the central nervous system can be guided — sometimes nudged, sometimes slammed — into wise policies. Put it to the vote of the body, in effect. ...When all goes well, harmony reigns and the various sources of wisdom in the body cooperate for the benefit of the whole, but we are all too familiar with the conflicts that can provoke the curious outburst "My body has a mind of its own!" Sometimes, apparently, it is tempting to lump together some of the embodied information into a separate mind. Why? Because it is organized in such a way that it can sometimes make independent discriminations, consult preferences, make decisions, enact policies that are in competition with your mind. At such time, the Cartesian perspective of a puppeteer self trying desperately to control an unruly body-puppet is very powerful. Your body can vigorously betray the secrets you are desperately trying to keep — by blushing and trembling or sweating, to mention only the most obvious cases. It can "decide" that in spite of your well-laid plans, right now would be a good time for sex, not intellectual discussion, and then take embarrassing steps in preparation for a coup d'etat. On another occasion, to your even greater chagrin and frustration, it can turn a deaf ear on your own efforts to enlist it for a sexual campaign, forcing you to raise the volume, twirl the dials, try all manner of preposterous cajolings to persuade it."

- Daniel Dennett

0 likesAcademics from the United StatesPhilosophers from the United StatesCognitive scientistsNon-fiction authors from the United StatesBiographers from the United States
"Since September 11, 2001, I have often thought that perhaps it was fortunate for the world that the attackers targeted the World Trade Center instead of the Statue of Liberty, for if they had destroyed our sacred symbol of democracy I fear we as Americans would have been unable to keep ourselves from indulging in paroxysms of revenge of a sort the world has never seen before. If that had happened, it would have befouled the meaning of the Statue of Liberty beyond any hope of subsequent redemption — if there were any people left to care. I have learned from my students that this upsetting thought of mine is subject to several unfortunate misconstruals, so let me expand on it to ward them off. The killing of thousands of innocents in the World Trade Center was a heinous crime, much more evil than the destruction of the Statue of Liberty would have been. And, yes, the World Trade Center was a much more appropriate symbol of al Qaeda's wrath than the Statue of Liberty would have been, but for that very reason it didn't mean as much, as a symbol, to us. It was Mammon and Plutocrats and Globalization, not Lady Liberty. I do suspect that the fury with which Americans would have responded to the unspeakable defilement of our cherished national symbol, the purest image of our aspirations as a democracy, would have made a sane and measured response extraordinarily difficult. This is the great danger of symbols — they can become too "sacred". An important task for religious people of all faiths in the twenty-first century will be spreading the conviction that there are no acts more dishonorable than harming "infidels" of one stripe or another for "disrespecting" a flag, a cross, a holy text."

- Daniel Dennett

0 likesAcademics from the United StatesPhilosophers from the United StatesCognitive scientistsNon-fiction authors from the United StatesBiographers from the United States
"[W]hat good to us is the gods' knowledge if we can't get it from them? How could one communicate with the gods? Our ancestors (while they were alive!) stumbled on an extremely ingenious solution: divination.We all know how hard it is to make the major decisions of life: should I hang tough or admit my transgression, should I move or stay in my present position, should I go to war or not, should I follow my heart or my head? We still haven't figured out any satisfactory systematic way of deciding these things. Anything that can relieve the burden of figuring out how to make these hard calls is bound to be an attractive idea.Consider flipping a coin, for instance. Why do we do it? To take away the burden of having to find a reason for choosing A over B. We like to have reasons for what we do, but sometimes nothing sufficiently persuasive comes to mind, and we recognize that we have to decide soon, so we concoct a little gadget, an external thing that will make the decision for us. But if the decision is about something momentous, like whether to go to war, or marry, or confess, anything like flipping a coin would be just too, well, flippant.In such a case, choosing for no good reason would be too obviously a sign of incompetence, and, besides, if the decision is really that important, once the coin has landed you'll have to confront the further choice: should you honor your just-avowed commitment to be bound by the flip of the coin, or should you reconsider? Faced with such quandaries, we recognize the need for some treatment stronger than a coin flip. Something more ceremonial, more impressive, like divination, which not only tells you what to do, but gives you a reason (if you squint just right and use your imagination).Scholars have uncovered a comically variegated profusion of ancient ways of delegating important decisions to uncontrollable externalities. Instead of flipping a coin, you can flip arrows (belomancy) or rods (rhabdomancy) or bones or cards (sortilege), and instead of looking at tea leaves (tasseography), you can examine the livers of sacrificed animals (hepatoscopy) or other entrails (haruspicy) or melted wax poured into water (ceroscopy). Then there is moleosophy (divination by blemishes), myomancy (divination by rodent behavior), nephomancy (divination by clouds), and of course the old favorites, numerology and astrology, among dozens of others."

- Daniel Dennett

0 likesAcademics from the United StatesPhilosophers from the United StatesCognitive scientistsNon-fiction authors from the United StatesBiographers from the United States
"The daily actions of religious people have accomplished uncounted good deeds throughout history, alleviating suffering, feeding the hungry, caring for the sick. Religions have brought the comfort of belonging and companionship to many who would otherwise have passed through this life all alone, without glory or adventure. They have not just provided first aid, in effect, for people in difficulties; they have provided the means for changing the world in ways that remove those difficulties. As Alan Wolfe says, "Religion can lead people out of cycles of poverty and dependency just as it led Moses out of Egypt". There is much for religion lovers to be proud of in their traditions, and much for all of us to be grateful for.The fact that so many people love their religions as much as, or more than, anything else in their lives is a weighty fact indeed. I am inclined to think that nothing could matter more than what people love. At any rate, I can think of no value that I would place higher. I would not want to live in a world without love. Would a world with peace, but without love, be a better world? Not if the peace was achieved by drugging the love (and hate) out of us, or by suppression. Would a world with justice and freedom, but without love, be a better world? Not if it was achieved by somehow turning us all into loveless law-abiders with none of the yearnings or envies or hatreds that are wellsprings of injustice and subjugation.It is hard to consider such hypotheticals, and I doubt if we should trust our first intuitions about them, but, for what it is worth, I surmise that we almost all want a world in which love, justice, freedom, and peace are all present, as much as possible, but if we had to give up one of these, it wouldn't — and shouldn't — be love. But, sad to say, even if it is true that nothing could matter more than love, it wouldn't follow from this that we don't have reason to question the things that we, and others, love. Love is blind, as they say, and because love is blind, it often leads to tragedy: to conflicts in which one love is pitted against another love, and something has to give, with suffering guaranteed in any resolution."

- Daniel Dennett

0 likesAcademics from the United StatesPhilosophers from the United StatesCognitive scientistsNon-fiction authors from the United StatesBiographers from the United States
"If I were designing a phony religion, I'd surely include a version of this little gem — but I'd have a hard time saying it with a straight face:If anybody ever raises questions of objections about our religion that you cannot answer, that person is almost certainly Satan. In fact, the more reasonable the person is, the more eager to engage you in open-minded and congenial discussion, the more sure you can be that you're talking to Satan in disguise! Turn away! Do not listen! It's a trap!What is particularly cute about this trick is that it is a perfect "wild card," so lacking in content that any sect or creed or conspiracy can use it effectively. Communist cells can be warned that any criticism they encounter is almost sure to be the work of FBI infiltrators in disguise, and radical feminist discussion groups can squelch any unanswerable criticism by declaring it to be phallocentric propaganda being unwittingly spread by a brainwashed dupe of the evil patriarchy, and so forth. This all-purpose loyalty-enforcer is paranoia in a pill, sure to keep the critics muted if not silent.Did anyone invent this brilliant adaptation, or is it a wild meme that domesticated itself by attaching itself to whatever memes were competing for hosts in its neighborhood? Nobody knows, but now it is available for anybody to use — although, if this book has any success, its virulence should diminish as people begin to recognize it for what it is."

- Daniel Dennett

0 likesAcademics from the United StatesPhilosophers from the United StatesCognitive scientistsNon-fiction authors from the United StatesBiographers from the United States
"Here is a well-known trajectory: You begin with a heartfelt desire to help other people and the conviction, however well or ill founded, that your guild or club or church is the coalition that can best serve to improve the welfare of others. If times are particularly tough, this conditional stewardship — I'm doing what's good for the guild because that will be good for everybody — may be displaced by the narrowest concern for the integrity of the guild itself, and for good reason: if you believe that the institution in question is the best path to goodness, the goal of preserving it for future projects, still unimagined, can be the most rational higher goal you can define. It is a short step from this to losing track of or even forgetting the larger purpose and devoting yourself singlemindedly to furthering the interests of the institution, at whatever costs. A conditional or instrumental allegiance can thus become indistinguishable in practice from a commitment to something "good in itself." A further short step perverts this parochial summum bonum to the more selfish goal of doing whatever it takes to keep yourself at the helm of the institution ("who better than I to lead us to triumph over our adversaries?")We have all seen this happen many times, and may even have caught ourselves in the act of forgetting just why we wanted to be leaders in the first place."

- Daniel Dennett

0 likesAcademics from the United StatesPhilosophers from the United StatesCognitive scientistsNon-fiction authors from the United StatesBiographers from the United States
"What [is] the prevailing attitude today among those who call themselves religious but vigorously advocate tolerance? There are three main options, ranging from the disingenuous Machiavellian--1. As a matter of political strategy, the time is not ripe for candid declarations of religious superiority, so we should temporize and let sleeping dogs lie in hopes that those of other faiths can gently be brought around over the centuries.--through truly tolerant Eisenhowerian "Our government makes no sense unless it is founded on a deeply religious belief — and I don't care what it is" --2. It really doesn't matter which religion you swear allegiance to, as long as you have some religion.--to the even milder Moynihanian benign neglect--3. Religion is just too dear to too many to think of discarding, even though it really doesn't do any good and is simply an empty historical legacy we can afford to maintain until it quietly extinguishes itself sometime in the distant and unforeseeable future.It it no use asking people which they choose, since both extremes are so undiplomatic we can predict in advance that most people will go for some version of ecumenical tolerance whether they believe it or not. ...We've got ourselves caught in a hypocrisy trap, and there is no clear path out. Are we like families in which the adults go through all the motions of believing in Santa Claus for the sake of the kids, and the kids all pretend still to believe in Santa Claus so as not to spoil the adults' fun? If only our current predicament were as innocuous and even comical as that! In the adult world of religion, people are dying and killing, with the moderates cowed into silence by the intransigence of the radicals in their own faiths, and many afraid to acknowledge what they actually believe for fear of breaking Granny's heart, or offending their neighbors to the point of getting run out of town, or worse.If this is the precious meaning our lives are vouchsafed thanks to our allegiance to one religion or another, it is not such a bargain, in my opinion. Is this the best we can do? Is it not tragic that so many people around the world find themselves enlisted against their will in a conspiracy of silence, either because they secretly believe that most of the world's population is wasting their lives in delusion (but they are too tenderhearted — or devious — to say so), or because they secretly believe that their own tradition is just such a delusion (but they fear for their own safety if they admit it)?"

- Daniel Dennett

0 likesAcademics from the United StatesPhilosophers from the United StatesCognitive scientistsNon-fiction authors from the United StatesBiographers from the United States
"Remember Marxism? It used to be a sour sort of fun to tease Marxists about the contradictions in some of their pet ideas. The revolution of the proletariat was inevitable, good Marxists believed, but if so, why were they so eager to enlist us in their cause? If it was going to happen anyway, it was going to happen with or without our help. But of course the inevitability that Marxists believe in is one that depends on the growth of the movement and all its political action. There were Marxists working very hard to bring about the revolution, and it was comforting to them to believe that their success was guaranteed in the long run. And some of them, the only ones that were really dangerous, believed so firmly in the rightness of their cause that they believed it was permissible to lie and deceive in order to further it. They even taught this to their children, from infancy. These are the "red-diaper babies," children of hardline members of the Communist Party of America, and some of them can still be found infecting the atmosphere of political action in left-wing circles, to the extreme frustration and annoyance of honest socialists and others on the left.Today we have a similar phenomenon brewing on the religious right: the inevitability of the End Days, or the Rapture, the coming Armageddon that will separate the blessed from the damned in the final day of Judgment. Cults and prophets proclaiming the imminent end of the world have been with us for several millennia, and it has been another sour sort of fun to ridicule them the morning after, when they discover that their calculations were a little off. But, just as with the Marxists, there are some among them who are working hard to "hasten the inevitable," not merely anticipating the End Days with joy in their hearts, but taking political action to bring about the conditions they think are the prerequisites for that occasion. And these people are not funny at all. They are dangerous, for the same reason that red-diaper babies are dangerous: they put their allegiance to their creed ahead of their commitment to democracy, to peace, to (earthly) justice — and to truth. If push comes to shove, some of the are prepared to lie and even to kill..."

- Daniel Dennett

0 likesAcademics from the United StatesPhilosophers from the United StatesCognitive scientistsNon-fiction authors from the United StatesBiographers from the United States
"Surely just about everybody has faced a moral dilemma and secretly wished, "If only somebody — somebody I trusted — could just tell me what to do!" Wouldn't this be morally inauthentic? Aren't we responsible for making our own moral decisions? Yes, but the virtues of "do it yourself" moral reasoning have their limits, and if you decide, after conscientious consideration, that your moral decision is to delegate further moral decisions in your life to a trusted expert, then you have made your own moral decision. You have decided to take advantage of the division of labor that civilization makes possible and get the help of expert specialists.We applaud the wisdom of this course in all other important areas of decision-making (don't try to be your own doctor, the lawyer who represents himself has a fool for a client, and so forth). Even in the case of political decisions, like which way to vote, the policy of delegation can be defended. … Is the a dereliction of [one's] dut[y] as a citizen? I don't think so, but it does depend on my having good grounds for trusting [the delegate's] judgment. … That why those who have an unquestioning faith in the correctness of the moral teachings of their religion are a problem: if they themselves haven't conscientiously considered, on their own, whether their pastors or priests or rabbis or imams are worthy of this delegated authority over their own lives, then they are in fact taking a personally immoral stand.This is perhaps the most shocking implication of my inquiry, and I do not shrink from it, even though it may offend many who think of themselves as deeply moral. It is commonly supposed that it is entirely exemplary to adopt the moral teachings of one's own religion without question, because -- to put it simply — it is the word of God (as interpreted, always, by the specialists to whom one has delegated authority). I am urging, on the contrary, that anybody who professes that a particular point of moral conviction is not discussable, not debatable, not negotiable, simply because it is the word of God, or because the Bible says so, or because "that is what all Muslims [Hindus, Sikhs ...] [sic] believe, and I am a Muslim [Hindu, Sikh ...]" [sic], should be seen to be making it impossible for the rest of us to take their views seriously, excusing themselves from the moral conversation, inadvertently acknowledging that their own views are not conscientiously maintained and deserve no further hearing."

- Daniel Dennett

0 likesAcademics from the United StatesPhilosophers from the United StatesCognitive scientistsNon-fiction authors from the United StatesBiographers from the United States
"One thing in particular struck me when I compared the medical world on which my life now depended with the religious institutions I have been studying so intensively in recent years. One of the gentler, more supportive themes to be found in every religion (so far as I know) is the idea that what really matters is what is in your heart: if you have good intentions, and are trying to do what (God says) is right, that is all anyone can ask. Not so in medicine! If you are wrong —especially if you should have known better — your good intentions count for almost nothing. And whereas taking a leap of faith and acting without further scrutiny of one's options is often celebrated by religions, it is considered a grave sin in medicine. A doctor whose devout faith in his personal revelations about how to treat aortic aneurysm led him to engage in untested trials with human patients would be severely reprimanded if not driven out of medicine altogether. There are exceptions, of course. A few swashbuckling, risk-taking pioneers are tolerated and (if they prove to be right) eventually honored, but they can exist only as rare exceptions to the ideal of the methodical investigator who scrupulously rules out alternative theories before putting his own into practice. Good intentions and inspiration are simply not enough.In other words, whereas religions may serve a benign purpose by letting many people feel comfortable with the level of morality they themselves can attain, no religion holds its members to the high standards of moral responsibility that the secular world of science and medicine does! And I'm not just talking about the standards 'at the top' — among the surgeons and doctors who make life or death decisions every day. I'm talking about the standards of conscientiousness endorsed by the lab technicians and meal preparers, too. This tradition puts its faith in the unlimited application of reason and empirical inquiry, checking and re-checking, and getting in the habit of asking "What if I'm wrong?" Appeals to faith or membership are never tolerated. Imagine the reception a scientist would get if he tried to suggest that others couldn't replicate his results because they just didn't share the faith of the people in his lab! And, to return to my main point, it is the goodness of this tradition of reason and open inquiry that I thank for my being alive today."

- Daniel Dennett

0 likesAcademics from the United StatesPhilosophers from the United StatesCognitive scientistsNon-fiction authors from the United StatesBiographers from the United States
"Goodness comes in many forms, not just medicine and science. Thank goodness for the music of, say, Randy Newman, which could not exist without all those wonderful pianos and recording studios, to say nothing of the musical contributions of every great composer from Bach through Wagner to Scott Joplin and the Beatles. Thank goodness for fresh drinking water in the tap, and food on our table. Thank goodness for fair elections and truthful journalism. If you want to express your gratitude to goodness, you can plant a tree, feed an orphan, buy books for schoolgirls in the Islamic world, or contribute in thousands of other ways to the manifest improvement of life on this planet now and in the near future. Or you can thank God — but the very idea of repaying God is ludicrous. What could an omniscient, omnipotent Being (the Man Who has Everything?) do with any paltry repayments from you? (And besides, according to the Christian tradition God has already redeemed the debt for all time, by sacrificing his own son. Try to repay that loan!) Yes, I know, those themes are not to be understood literally; they are symbolic. I grant it, but then the idea that by thanking God you are actually doing some good has got to be understood to be just symbolic, too. I prefer real good to symbolic good. Still, I excuse those who pray for me. I see them as like tenacious scientists who resist the evidence for theories they don't like long after a graceful concession would have been the appropriate response. I applaud you for your loyalty to your own position — but remember: loyalty to tradition is not enough. You've got to keep asking yourself: What if I'm wrong? In the long run, I think religious people can be asked to live up to the same moral standards as secular people in science and medicine."

- Daniel Dennett

0 likesAcademics from the United StatesPhilosophers from the United StatesCognitive scientistsNon-fiction authors from the United StatesBiographers from the United States
"Among the many debates over the origins and evolution of war is whether humans have been getting more violent or less. Steven Pinker and others who think like him, such as the archaeologist Ian Morris, are on the optimistic side and believe there is a clear trend away from violence. Most countries no longer have public executions; they have laws against cruelty to animals or children; and sports such as bear baiting or dog fighting are normally illegal. The optimists go further and attempt to tot up deaths from war in the past – not in itself an easy task – to argue that homicide rates in the past were far higher than they are today and that deaths in war, as a proportion of humans alive at the time, are fewer in the twentieth and twenty-first centuries, even allowing for the great bloodlettings of the two world wars, than in earlier wars. Others challenge the figures and also point out that war deaths in the twentieth century may amount to 75 per cent of all war deaths in the last 5,000 years. And if you really want to be depressed about the prospects for humanity, studies using mathematical tools at the University of Florence and the University of Colorado claim to show that the trend is for fewer but more deadly wars. Their argument is that the more interconnected societies become the quicker a conflict can spread along the paths of the network – just as computer viruses or forest fires do. A small squabble in the Balkans in the summer of 1914 grew into the Great War because Europe’s powers were so interlinked by treaties, understandings or plans that tensions spread upwards and outwards from the assassination of the Archduke Franz Ferdinand in Sarajevo until a general war exploded. Even if Pinker is right – and the debate goes on – it somehow does not seem very reassuring. Those of us who have enjoyed the Long Peace since 1945 need to reflect that much of the world, including Indochina, Afghanistan, the Great Lakes district in Africa and large parts of the Middle East, has seen and still sees conflict. A long-running project at Uppsala University in Sweden estimates that between 1989 and 2017 over 2 million people died as a result of war. Since 1945 perhaps 52 million have been forced to flee because of conflict."

- Steven Pinker

0 likesAcademics from CanadaLinguists from CanadaCognitive scientistsPsychologists from CanadaPhilosophers from Canada
"There are no really solipsistic philosophers, and those who think they are deceive themselves. The true solipsist feels at one with the universe, and so very identical to it that he does not even feel the need for two terms. The true solipsist projects all his states of mind onto things. The true solipsist is entirely alone in the world, that is, he has no notion of anything exterior to himself. In other words the true solipsist has no idea of self. There is no self: there is the world. It is in this sense it is reasonable to call a baby a solipsist: the feelings and desires of a baby know no limits since they are a part of everything he sees, touches, and perceives. Babies are, then, obviously narcissistic, but not in the way adults are, not even Spinoza's God, and I am a little afraid that Freud sometimes forgets that the narcissistic baby has no sense of self. Given this definition of solipsism, egocentrism in children clearly appears to be a simple continuation of solipsism in infants.. Egocentrism, as we have seen, is not an intentional or even a conscious process. A child has no idea that he is egocentric. He believes everybody thinks the way he does, and this false universality is due simply to an absence of the sense of limits on his individuality. In this light, egocentrism and solipsism are quite comparable: both stem from the absence or the weakness of the sense of self."

- Jean Piaget

0 likesAcademics from SwitzerlandPhilosophers from SwitzerlandCognitive scientistsPsychologists from SwitzerlandScience authors
"From the point of view of the practice or application of rules four successive stages can be distinguished. A first stage of a purely motor and individual character, during which the child handles the marbles at the dictation of his desires and motor habits. This leads to the formation of more or less ritualized schemas, but since play is still purely individual, one can only talk of motor rules and not of truly collective rules. The second may be called egocentric for the following reasons. This stage begins at the moment when the child receives from outside the example of codified rules, that is to say, some time between the ages of two and five. But though the child imitates this example, he continues to play either by himself without bothering to find play-fellows, or with others, but without trying to win, and therefore without attempting to unify the different ways of playing. In other words, children of this stage, even when they are playing together, play each one "on his own " (everyone can win at once) and without regard for any codification of rules. This dual character, combining imitation of others with a purely individual use of the examples received, we have designated by the term Egocentrism. A third stage appears between 7 and 8, which we shall call the stage of incipient cooperation. Each player now tries to win, and all, therefore, begin to concern themselves with the question of mutual control and of unification of the rules. But while a certain agreement may be reached in the course of one game, ideas about the rules in general are still rather vague. In other words, children of 7-8, who belong to the same class at school and are therefore constantly playing with each other, give, when they are questioned separately, disparate and often entirely contradictory accounts of the rules observed in playing marbles. Finally, between the years of 11 and 12, appears a fourth stage, which is that of the codification of rules. Not only is every detail of procedure in the game fixed, but the actual code of rules to be observed is known to the whole society. There is remarkable concordance in the information given by children of 10-12 belonging to the same class at school, when they are questioned on the rules of the game and their possible variations."

- Jean Piaget

0 likesAcademics from SwitzerlandPhilosophers from SwitzerlandCognitive scientistsPsychologists from SwitzerlandScience authors
"A second prefatory question faces us: that of society and the individual. We have sought to contrast the child and the civilized adult on the ground of their respective social attitudes. The baby (at the stage of motor intelligence) is asocial, the egocentric child is subject to external constraint but has little capacity for cooperation, the civilized adult of to-day presents the essential character of cooperation between differentiated personalities who regard each other as equals. There are therefore three types of behavior: motor behavior, egocentric behavior (with external constraint), and cooperation. And to these three types of social behavior there correspond three types of rules: motor rules, rules due to unilateral respect, and rules due to mutual respect. But here again, one must beware of laying down the law: for things are motor, individual and social all at once. As we shall have occasion to show, rules of cooperation are in some respects the outcome of the rules of coercion and of the motor rules. On the other hand, coercion is applied during the first days of an infant's life, and the earliest social relations contain the germs of cooperation. Here again, it is not so much a question of these successive features themselves as of the proportions in which they are present. Moreover, the way in which conscious realization and the time-lag from one level to another come into play is a further bar to our arranging these phenomena in a strict sequence, as though they made a single appearance and then disappeared from the scene once and for all."

- Jean Piaget

0 likesAcademics from SwitzerlandPhilosophers from SwitzerlandCognitive scientistsPsychologists from SwitzerlandScience authors
"Every observer has noted that the younger the child, the less sense he has of his own ego. From the intellectual point of view, he does not distinguish between external and internal, subjective and objective. From the point of view of action, he yields to every suggestion, and if he does oppose to other people's wills — a certain negativism which has been called "the spirit of contradiction" — this only points to his real defenselessness against his surroundings. A strong personality can maintain itself without the help of this particular weapon. The adult and the older child have complete power over him. They impose their opinions and their wishes, and the child accepts them without knowing that he does so. Only — and this is the other side of the picture — as the child does not dissociate his ego from the environment, whether physical or social, he mixes into all his thoughts and all his actions, ideas and practices that are due to the intervention of his ego and which, just because he fails to recognize them as subjective, exercise a check upon his complete socialization. From the intellectual point of view, he mingles his own fantasies with accepted opinions, whence arise pseudo lies (or sincere lies), syncretism, and all the features of child thought. From the point of view of action, he interprets in his own fashion the examples he has adopted, whence the egocentric form of play we were examining above. The only way of avoiding these individual refractions would lie in true cooperation, such that both child and senior would each make allowance for his own individuality and for the realities that were held in common."

- Jean Piaget

0 likesAcademics from SwitzerlandPhilosophers from SwitzerlandCognitive scientistsPsychologists from SwitzerlandScience authors
"It is when the child is accustomed to act from the point of view of those around him, when he tries to please rather than to obey, that he will judge in terms of intentions. So that taking intentions into account presupposes cooperation and mutual respect. Only those who have children of their own know how difficult it is to put this into practice. Such is the prestige of parents in the eyes of the very young child, that even if they lay down nothing in the form of general duties, their wishes act as law and thus give rise automatically to moral realism (independently, of course, of the manner in which the child eventually carries out these desires). In order to remove all traces of moral realism, one must place oneself on the child's own level, and give him a feeling of equality by laying stress on one's own obligations and one's own deficiencies. In this way the child will find himself in the presence, not of a system of commands requiring ritualistic and external obedience, but of a system of social relations such that everyone does his best to obey the same obligations, and does so out of mutual respect. The passage from obedience to cooperation thus marks a progress analogous to that of which we saw the effects in the evolution of the game of marbles: only in the final stage does the morality of intention triumph over the morality of objective responsibility. When parents do not trouble about such considerations as these, when they issue contradictory commands and are inconsistent in the punishments they inflict, then, obviously, it is not because of moral constraint but in spite of and as a reaction against it that the concern with intentions develops in the child. Here is a child, who, in his desire to please, happens to break something and is snubbed for his pains, or who in general sees his actions judged otherwise than he judges them himself. It is obvious that after more or less brief periods of submission, during which he accepts every verdict, even those that are wrong, he will begin to feel the injustice of it all. Such situations can lead to revolt. But if, on the contrary, the child finds in his brothers and sisters or in his playmates a form of society which develops his desire for cooperation and mutual sympathy, then a new type of morality will be created in him, a morality of reciprocity and not of obedience. This is the true morality of intention and of subjective responsibility."

- Jean Piaget

0 likesAcademics from SwitzerlandPhilosophers from SwitzerlandCognitive scientistsPsychologists from SwitzerlandScience authors
"Every individual word in a passage or poetry can no more be said to denote some specific referent than does every brush mark, every line in a painting have its counterpart in reality. The writer or speaker does not communicate his thoughts to us; he communicates a representation for carrying out, this function under the severe discipline of using the only materials he has, sound and gesture. Speech is like painting, a representation made out of given materials -- sound or paint. The function of speech is to stimulate and set up thoughts in us having correspondence with the speaker's desires; he has then communicated with us. But he has not transmitted a copy of his thoughts, a photograph, but only a stream of speech -- a substitute made from the unpromising material of sound. The artist, the sculptor, the caricaturist, the composer are akin in this [fact that they have not transmitted a copy of their thoughts], that they express (make representations of) their thoughts using chosen, limited materials. They make the "best" representations, within these self-imposed constraints. A child who builds models of a house, or a train, using only a few colored bricks, is essentially engaged in the same creative task.* Metaphors can play a most forceful role, by importing ideas through a vehicle language, setting up what are purely linguistic associations (we speak of "heavy burden of taxation," "being in a rut"). The imported concepts are, to some extent, artificial in their contexts, and they are by no means universal among different cultures. For instance, the concepts of cleanliness and washing are used within Christendom to imply "freedom from sin." We Westerners speak of the mind's eye, but this idea is unknown amongst the Chinese. After continued use, many metaphorical words become incorporated into the language and lose their original significance; words such as "explain," "ponder," "see (what you mean)" we no longer think of as metaphorical. Metaphors arise because we continually need to stretch the range of words as we accumulate new concepts and abstract relationships. A printed text is not simply a chain of individual words, picked one at a time; it is a whole. It has a structure, but it has meaning for us only if it represents a continuity of our experience of past texts. A text in some strange foreign language set up an abrupt change in our experience, a discontinuity, and we make nothing of it. Given a translator's dictionary we may decipher some of the words and attain some understanding, though this understanding through translation has been achieved by projecting the text onto our own language; that is, we are looking at it with the eyes of our English-speaking culture. A grammar book may help us to decipher the text more thoroughly, and help us comprehend something of the language structure, but we may never fully understand if we are not bred in the culture and society that has modeled and shaped the language. (p. 74)"

- Colin Cherry

0 likesAcademics from the United KingdomCognitive scientistsPeople from London
""Information" in most, if not all, of its connotations seems to rest upon the notion of selective power. The Shannon theory regards the information source, in emitting the signals (signs), as exerting a selective power upon the ensemble of messages. In the Carnap-Bar-Hillel semantic theory, the information content of statements relates to the selective power they exert upon ensembles of states. Again, at its pragmatic level, in true communicative situations [...] a source of information has a certain value to a recipient, where "value" may be regarded as a "selective power." Gabor, for example, observes that what people value in a source of information (i.e., what they are prepared to pay for) depends upon its exclusiveness and prediction power; he cites instances of a newspaper editor hoping for a "scoop" and a racegoer receiving information from a tipster. "Exclusiveness" here implies the selecting of that one particular recipient out of the population, while the "prediction" value of information rests upon the power it gives to the recipient to select his future action, out of the whole range of prior uncertainty as to what action to take. Again, signs have the power to select responses in people, such responses depending upon a totality of conditions. Human communication channels consist of individuals in conversation, or in various forms of social intercourse. Each individual and each conversation is unique; different people react to signs in different ways, depending each upon their own past experiences and upon the environment at the time. It is such variations, such differences, which gives rise to the principal problems in the study of human communication. (p. 244-5)"

- Colin Cherry

0 likesAcademics from the United KingdomCognitive scientistsPeople from London
"As we survey the various stages of evolution, from the simplest one-cell creatures up to man. we see a steady improvement in the methods of learning and adaptation to a hostile world. Each step in learning ability gives better adaptation and greater chance of survival. We are carried a long way up the scale by innate reflexes and rudimentary muscular learning faculties. Habits indeed, not rational thought, assist us to surmount most of life's obstacles. Most, but by no means all; for learning in the high mammals exhibits the unexplained phenomenon of "insight," which shows itself by sudden changes in behavior in learning situations -- in sudden departures from one method of organizing a task, or solving a problem, to another. Insight, expectancy, set, are the essentially "mind-like" attributes of communication, and it is these, together with the representation of concepts, which require physiological explanation. At the higher end of the scale of evolution, this quality we call "mind" appears more and more prominently, but it is at our own level that learning of a radically new type has developed -- through our powers of organizing thoughts, comparing and setting them into relationship, especially with the use of language. We have a remarkable faculty of forming generalizations, of recognizing universals, of associating and developing them. It is our multitude of general concepts, and our powers of organizing them with the aid of language in varied ways, which forms the backbone of human communication, and which distinguishes us from the animals. (p. 304)"

- Colin Cherry

0 likesAcademics from the United KingdomCognitive scientistsPeople from London
"Science traditionally takes the reductionist approach, saying that the collective properties of molecules, or the fundamental units of whatever system you're talking about, are enough to account for all of the system's activity. But this standard approach leaves out one very important additional factor, and that's the spacing and timing of activity — its pattern or form. The components of any system are linked up in different ways, and these possible relationships, especially at the higher levels, are not completely covered by the physical laws for the elementary interactions between atoms and molecules. At some point, the higher properties of the whole begin to take over and govern the fate of its constituents. A simple way to illustrate this idea is to imagine a molecule in an airplane flying from L.A. to New York. The molecule may be jostled somewhat or held in position by its neighbors, but these lower-level actions are trivial compared to its movement as the plane flies across the continent. If you plot the movement of the molecule through time and space, those features governed by the higher properties of the plane as a whole make those controlled at the level of the molecule insignificant by comparison. The higher properties control the lower, not by direct intervention, but by supervention."

- Roger Wolcott Sperry

0 likesAcademics from the United StatesBiologists from the United StatesPsychologists from the United StatesCognitive scientistsNeuroscientists
"People who had damage to the right cerebral hemisphere were unable to recognise simple patterns, or enjoy music, but they could still speak normally. People with left-brain damage were able to recognise patterns, but their speech was impaired. Obviously, then, the left deals with language, and you would expect a split-brain patient to be unable to read with his right eye (connected, remember, to the opposite side of the brain). Sperry's patient was also unable to write anything meaningful (i.e., complicated) with his left hand. They noticed another oddity. if the patient bumped into something with his left side, he did not notice. And the implications were very odd indeed. Not only did the split-brain operation give the patient two separate minds; it also seemed to restrict his identity, or ego, to the left side. When they placed an object in his left hand, and asked him what he was holding, he had no idea. Further experiments underlined the point. If a split-brain patient is shown two different symbols -- say a circle and a square -- with each eye, and is asked to say what he has just seen, he replies, 'A square'. Asked to draw with his left hand what he has seen, and he draws a circle. Asked what he has just drawn, he replies: 'A square'. And when one split-brain patient was shown a picture of a nude male with the right-brain, she blushed; asked why she was blushing, she replied truthfully: 'I don't know'. The implications are clearly staggering. The person you call 'you' lives in the left side of your brain. And a few centimeters away there is another person, a completely independent identity. Where language is concerned, this other person is almost an imbecile. In other respects, he is more competent than the inhabitant of the left-brain; for example, he can make a far more accurate perspective drawing of a house. In effect, the left-brain person is a scientist, the right-brain an artist."

- Roger Wolcott Sperry

0 likesAcademics from the United StatesBiologists from the United StatesPsychologists from the United StatesCognitive scientistsNeuroscientists
"I have not attempted in these papers to put forward any grand view of the nature of philosophy; nor do I have any such grand view to put forward if I would. It will be obvious that I do not agree with those who see philosophy as the history of 'howlers', and progress in philosophy as the debunking of howlers. It will also be obvious that I do not agree with those who see philosophy as the enterprise of putting forward a priori truths about the real world (since, for one thing, there are no a priori truths, in my view). I see philosophy as a field which has certain central questions, for example, the relation between thought and reality, and, to mention some questions about which I have not written, the relation between freedom and responsibility, and the nature of the good life. It seems obvious that in dealing with these questions philosophers have formulated rival research programs, that they have put forward general hypotheses, and that philosophers within each major research program have modified their hypotheses by trial and error, even if they sometimes refuse to admit that that is what they are doing. To that extent philosophy is a 'science'. To argue about whether philosophy is a science in any more serious sense seems to me to be hardly a useful occupation. The important thing is that in spite of the stereotypes of science and philosophy that have become blinkers inhibiting the view of laymen, scientists, and philosophers, science and philosophy are interdependent activities; philosophers have always found it essential to draw upon the scientific knowledge of the time, and scientists have always found it essential to do a certain amount of philosophy in their very scientific work, even if they denied that that was what they were doing. It does not seem to me important to decide whether science is philosophy or philosophy is science as long as one has a conception of both that makes both essential to a responsible view of the real world and of man's place in it."

- Hilary Putnam

0 likesWomen academics from the United StatesPrinceton University facultyCognitive scientistsComputer scientists from the United StatesMathematicians from the United States
"I do not agree with Quine, that there is no analytic-synthetic distinction to be drawn at all. But I do believe that his emphasis on the monolithic character of our conceptual system and his negative emphasis on the silliness of regarding mathematics as consisting in some sense of 'rules of language', represent exceedingly important theoretical insights in philosophy. I think that what we have to do now is to settle the relatively trivial question concerning analytic statements properly so called ('All bachelors are unmarried'). We have to take a fresh look at the framework principles so much discussed by philosophers, disabusing ourselves of the idea that they are 'rules of language' in any literal or lexicographic sense; and above all, we have to take a fresh look at the nature of logical and mathematical truths. With Quine's contribution, we have to face two choices: We can ignore it and go on talking about the 'logic' of individual words. In that direction lies sterility and more, much more, of what we have already read. The other alternative is to face and explore the insight achieved by Quine, trying to reconcile the fact that Quine is overwhelmingly right in his critique of what other philosophers have done with the analytic-synthetic distinction with the fact that Quine is wrong in his literal thesis, namely, that the distinction itself does not exist at all. In the latter direction lies philosophic progress. For philosophic progress is nothing if it is not the discovery of new areas for dialectical exploration."

- Hilary Putnam

0 likesWomen academics from the United StatesPrinceton University facultyCognitive scientistsComputer scientists from the United StatesMathematicians from the United States