• Refine Query
  • Source
  • Publication year
  • to
  • Language
  • 88
  • 69
  • 60
  • 16
  • 6
  • 5
  • 4
  • 4
  • 4
  • 2
  • 1
  • 1
  • 1
  • Tagged with
  • 417
  • 417
  • 114
  • 92
  • 92
  • 92
  • 92
  • 90
  • 70
  • 59
  • 58
  • 54
  • 54
  • 53
  • 52
  • About
  • The Global ETD Search service is a free service for researchers to find electronic theses and dissertations. This service is provided by the Networked Digital Library of Theses and Dissertations.
    Our metadata is collected from universities around the world. If you manage a university/consortium/country archive and want to be added, details can be found on the NDLTD website.
1

Between Modern and Postmodern: Nietzsche on Truth and Knowledge

Laderoute, Karl W. 10 1900 (has links)
<p>This thesis examines Nietzsche’s epistemology. Its main interlocutors are two previously existing attempts to explain Nietzsche’s views on truth and knowledge. One of these interpretationsI dub the ‘postmodern’ reading, held most notably by Sarah Kofman, Jacques Derrida, and Paul de Man. The other is the ‘modern’ reading of Walter Kaufmann, John T. Wilcox, and most prominently Maudemarie Clark and Brian Leiter. Each of these readings emphasizes one aspect of Nietzsche’s thought. The postmodern reading focuses on Nietzsche’s more radical pronouncements, and promotes a type of scepticism and subjectivism. The modern reading, by contrast, emphasizes Nietzsche’s more traditional claims, and argues that he lauds science and preserves our ability to attain truth. However, neither reading is entirely satisfactory. In what follows, I first critically examine both of these readings in detail. The first chapter highlights the major points of these two readings, as well as some issues in each. After detailing these positions, I then turn to a largely chronological reading of Nietzsche’s works to establish an alternative account of his epistemology. Chapters two through four provide readings of Nietzsche’s epistemological claims in his major works from Human, All Too Human (1878) until Twilight of the Idols(1888). I combine this chronological reading withother informative aspects of Nietzsche’s thought. These other aspects include Nietzsche’s reading of Roger Boscovich (1711-1787) and his adoption of force-point ontology, his ontological commitment to nominalism, his views on evolution and its role in epistemology, and his similarities with Ernst Mach (1838-1916). Finally, I also connect Nietzsche’s epistemology with his critiques of morality and religion. I show that my reading is buttressed by the deep congruity between Nietzsche’s epistemology and his critiques, while the modern and postmodern readings are both unable to account for this congruity in a satisfactory manner.</p> / Doctor of Philosophy (PhD)
2

Hegel and Spinoza

Stephenson, A. H. January 1985 (has links)
No description available.
3

The power of interests in early-modern English political thought

Karstadt, Elliott January 2013 (has links)
This thesis studies the relationship between the particular interests of individuals and the common good, as it is conceived by various moral and political philosophers in earlymodern England (c.1640-c.1740). Interests are spoken of in English translations of Italian and French texts in the early seventeenth century, and are often used to describe goods or desires that are morally ambiguous. The vocabulary becomes commonly used in political tracts during the English Civil Wars, and this is where the thesis begins. We then move on to an analysis of the place of interests in Hobbes’s changing civil science. Hobbes continues to see interests as being morally ambiguous and dangerous to the common good. The third chapter deals with the republican tradition (epitomized by James Harrington), in which thinkers begin to conceive how interests might be manipulated to serve the common good. Chapter 4 deals with the men of latitude of the Restoration, who first conceive that interests are in fact identical with our moral virtues. We thereby come to see that the important questions regarding interests in the restoration revolved around religion and morality, rather than (as is commonly assumed) around trade. The fifth chapter deals with the commonwealth theorists, who became increasingly concerned that Charles II’s court, and subsequently the court whigs, were beginning to constitute an interest separate from that of the people. We then come to a discussion of Bernard Mandeville, who is generally thought to be a critic of the commonwealthmen, but (in his use of the vocabulary of interests) actually bears quite a close intellectual resemblance to them. The thesis ends with an account of a number of reactions to Mandeville, chief among them, Joseph Butler, who argues that not only are our interests identical with virtue, but they also naturally serve the common good.
4

Some antecedents of Leibniz's principles

Castro, Martinho Antonio Bittencourt de, History & Philosophy, Faculty of Arts & Social Sciences, UNSW January 2008 (has links)
Leibniz considered that scepticism and confusion engendered by the disputes of different sects or schools of metaphysics were obstacles to the progress of knowledge in philosophy. His solution was to adopt an eclectic method with the aim of uncovering the truth hidden beneath the dispute of schools. Leibniz's project was, having in mind the eclectic method, to synthesise a union between old pre-modern philosophy, based on formal and final causes, and new modern philosophy which gave preference to efficient causes. The result of his efforts is summarised in the Monadology. But the question remains: to what degree was Leibniz successful in this enterprise? An objective of this thesis is to investigate whether philosophical tradition can justify or support some of the arguments that are at the basis of Leibniz's system (for example, monads have no window to the exterior world, a phrase that summarises the structure of Monadology). I shall demonstrate how Leibniz reflects the concerns and the positions of his key predecessors. Thus, the aim of the thesis is to explore key antecedents to Leibniz's central doctrines. The thesis argues that Leibniz carried out a logical development of some principles in the eclectic system of Plotinus, which resulted in the structure of Monadology. Whether Leibniz's project was successful or not will depend on how we view those principles.
5

How to teach an old dog new tricks:quantum information, quantum computing, and the philosophy of physics

Duwell, Armond James 31 January 2005 (has links)
My dissertation consists of two independent parts. Part one of my dissertation examines concepts of quantum information. I clarify three very different concepts of information and assess their implications for understanding quantum mechanics. First I clarify the concept of information due to Shannon, and its relation to physical theories. Using the Shannon concept, I examine two purportedly new concepts of quantum information. I argue that a fundamental philosophical mistake is made regarding these concepts. Advocates of these new concepts do not properly distinguish between the properties of information due to the physical medium it is stored in from the properties of information per se. This distinction is crucial for developing a new concept to help us understand quantum mechanics and evaluating its merits. Part two of my dissertation examines explanations of the efficiency that quantum computers enjoy over classical computers for some computational tasks, and the relationship between explanations of efficiency and interpretations of quantum mechanics. I examine the so-called quantum parallelism thesis, that quantum computers can perform many computations in a single step, a feat thought not to be possible on classical computers. The truth of this thesis is not obvious, and contested by some. I develop a set of general criteria for computation that any computing device must satisfy. I use these criteria to demonstrate that the quantum parallelism thesis is true. As an application of these general criteria for computation I articulate three distinct concepts of parallelism and demonstrate that classical computers can compute in parallel as well. This demonstrates that the truth of the quantum parallelism thesis alone does not provide a complete explanation of the efficiency of quantum computers. I supplement the quantum parallelism thesis to provide a complete explanation. Finally, I address the claim that only the many-worlds interpretation of quantum mechanics can underwrite the truth of the quantum parallelism thesis. The general criteria for computation provide support for the quantum parallelism thesis independent of any interpretation of quantum mechanics.
6

Empiricism and the Epistemic Status of Imaging Technologies

Delehanty, Megan Catherine 04 October 2005 (has links)
This starting point for this project was the question of how to understand the epistemic status of mathematized imaging technologies such as positron emission tomography (PET) and confocal microscopy. These sorts of instruments play an increasingly important role in virtually all areas of biology and medicine. Some of these technologies have been widely celebrated as having revolutionized various fields of studies while others have been the target of substantial criticism. Thus, it is essential that we be able to assess these sorts of technologies as methods of producing evidence. They differ from one another in many respects, but one feature they all have in common is the use of multiple layers of statistical and mathematical processing that are essential to data production. This feature alone means that they do not fit neatly into any standard empiricist account of evidence. Yet this failure to be accommodated by philosophical accounts of good evidence does not indicate a general inadequacy on their part since, by many measures, they very often produce very high quality evidence. In order to understand how they can do so, we must look more closely at old philosophical questions concerning the role of experience and observation in acquiring knowledge about the external world. Doing so leads us to a new, grounded version of empiricism. After distinguishing between a weaker and a stronger, anthropomorphic version of empiricism, I argue that most contemporary accounts of observation are what I call benchmark strategies that, implicitly or explicitly, rely on the stronger version according to which human sense experience holds a place of unique privilege. They attempt to extend the bounds of observation and the epistemic privilege accorded to it by establishing some type of relevant similarity to the benchmark of human perception. These accounts fail because they are unable to establish an epistemically motivated account of what relevant similarity consists of. The last best chance for any benchmark approach, and, indeed, for anthropomorphic empiricism, is to supplement a benchmark strategy with a grounding strategy. Toward this end, I examine the Grounded Benchmark Criterion which defines relevant similarity to human perception to be defined in terms of the reliability-making features of human perception. This account, too, must fail due to our inability to specify these features given the current state of understanding of the human visual system. However, this failure reveals that it is reliability alone that is epistemically relevant, not any other sort of similarity to human perception. Current accounts of reliability suffer from a number of difficulties, so I develop a novel account of reliability that is based on the concept of granularity. My account of reliability in terms of a granularity match both provides the means to refine the weaker version of empiricism and allows us to establish when and why imaging technologies are reliable. Finally, I use this account of granularity is examining the importance of the fact that the output of imaging technologies usually is images.
7

Explicating Emotions

Scarantino, Andrea 10 October 2005 (has links)
In the course of their long intellectual history, emotions have been identified with items as diverse as perceptions of bodily changes (feeling tradition), judgments (cognitivist tradition), behavioral predispositions (behaviorist tradition), biologically based solutions to fundamental life tasks (evolutionary tradition), and culturally specific social artifacts (social constructionist tradition). The first objective of my work is to put some order in the mare magnum of theories of emotions. I taxonomize them into families and explore the historical origin and current credentials of the arguments and intuitions supporting them. I then evaluate the methodology of past and present emotion theory, defending a bleak conclusion: a great many emotion theorists ask What is an emotion? without a clear understanding of what counts as getting the answer right. I argue that there are two ways of getting the answer right. One is to capture the conditions of application of the folk term "emotion" in ordinary language (Folk Emotion Project), and the other is to formulate a fruitful explication of it (Explicating Emotion Project). Once we get clear on the desiderata of these two projects, we realize that several long-running debates in emotion theory are motivated by methodological confusions. The constructive part of my work is devoted to formulating a new explication of emotion suitable for the theoretical purposes of scientific psychology. At the heart of the Urgency Management System (UMS) theory of emotions I propose is the idea that an umotion is a special type of superordinate system which instantiates and manages an urgent action tendency by coordinating the operation of a cluster of cognitive, perceptual and motoric subsystems. Crucially, such superordinate system has a proper function by virtue of which it acquires a special kind of intentionality I call pragmatic. I argue that umotion is sufficiently similar in use to emotion to count as explicating it, it has precise rules of application, and it accommodates a number of central and widely shared intuitions about the emotions. My hope is that future emotion research will demonstrate the heuristic fruitfulness of the umotion concept for the sciences of mind.
8

Subjective Measures of Well-Being: A philosophical examination

Angner, Erik 30 September 2005 (has links)
Over the last couple of decades, as part of the rise of positive psychology, psychologists have given increasing amounts of attention to so-called subjective measures of well-being. These measures, which are supposed to represent the well-being of individuals and groups, are often presented as alternatives to more traditional economic ones for purposes of the articulation, implementation and evaluation of public policy. Unlike economic measures, which are typically based on data about income, market transactions and the like, subjective measures are based on answers to questions like: Taking things all together, how would you say things are these days would you say youre very happy, pretty happy, or not too happy these days? The aim of this dissertation is to explore issues in the philosophical foundations of subjective measures of well-being, with special emphasis on the manner in which the philosophical foundations of subjective measures differ from those of traditional economic measures. Moreover, the goal is to examine some arguments for and against these measures, and, in particular, arguments that purport to demonstrate the superiority of economic measures for purposes of public policy. My main thesis is that the claim that subjective measures of well-being cannot be shown to be inferior to economic measures quite as easily as some have suggested, but that they nevertheless are associated with serious problems, and that questions about the relative advantage of subjective and economic measures for purposes of public policy will depend on some fundamentally philosophical judgments, e.g. about the nature of well-being and the legitimate goals for public policy.
9

Computations and Computers in the Sciences of Mind and Brain

Piccinini, Gualtiero 18 November 2003 (has links)
Computationalism says that brains are computing mechanisms, that is, mechanisms that perform computations. At present, there is no consensus on how to formulate computationalism precisely or adjudicate the dispute between computationalism and its foes, or between different versions of computationalism. An important reason for the current impasse is the lack of a satisfactory philosophical account of computing mechanisms. The main goal of this dissertation is to offer such an account. I also believe that the history of computationalism sheds light on the current debate. By tracing different versions of computationalism to their common historical origin, we can see how the current divisions originated and understand their motivation. Reconstructing debates over computationalism in the context of their own intellectual history can contribute to philosophical progress on the relation between brains and computing mechanisms and help determine how brains and computing mechanisms are alike, and how they differ. Accordingly, my dissertation is divided into a historical part, which traces the early history of computationalism up to 1946, and a philosophical part, which offers an account of computing mechanisms. The two main ideas developed in this dissertation are that (1) computational states are to be identified functionally not semantically, and (2) computing mechanisms are to be studied by functional analysis. The resulting account of computing mechanism, which I call the functional account of computing mechanisms, can be used to identify computing mechanisms and the functions they compute. I use the functional account of computing mechanisms to taxonomize computing mechanisms based on their different computing power, and I use this taxonomy of computing mechanisms to taxonomize different versions of computationalism based on the functional properties that they ascribe to brains. By doing so, I begin to tease out empirically testable statements about the functional organization of the brain that different versions of computationalism are committed to. I submit that when computationalism is reformulated in the more explicit and precise way I propose, the disputes about computationalism can be adjudicated on the grounds of empirical evidence from neuroscience.
10

Representations of Space in Seventeenth Century Physics

Miller, David Marshall 02 June 2006 (has links)
The changing understanding of the universe that characterized the birth of modern science included a fundamental shift in the prevailing representation of space the presupposed conceptual structure that allows one to intelligibly describe the spatial properties of physical phenomena. At the beginning of the seventeenth century, the prevailing representation of space was spherical. Natural philosophers first assumed a spatial center, then specified meanings with reference to that center. Directions, for example, were described in relation to the center, and locations were specified by distance from the center. Through a series of attempts to solve problems first raised by the work of Copernicus, this Aristotelian, spherical framework was replaced by a rectilinear representation of space. By the end of the seventeenth century, descriptions were understood by reference to linear orientations, as parallel or oblique to a presupposed line, and locations were identified without reference to a privileged central point. This move to rectilinear representations of space enabled Gilbert, Kepler, Galileo, Descartes, and Newton to describe and explain the behavior of the physical world in the novel ways for which these men are justly famous, including their theories of gravitational attraction and inertia. In other words, the shift towards a rectilinear representation of space was essential to the fundamental reconception of the universe that gave rise to both modern physical theory and, at the same time, the linear way of experiencing the world that characterizes modern science.

Page generated in 0.14 seconds