Literally just mainlining marketing material straight into whatever’s left of their rotting brains.

    • PolandIsAStateOfMind@lemmygrad.ml
      link
      fedilink
      English
      arrow-up
      7
      ·
      edit-2
      1 year ago

      You’re right that it isn’t, though considering science have huge problems even defining sentience, it’s pretty moot point right now. At least until it start to dream about electric sheep or something.

      • UlyssesT [he/him]@hexbear.net
        link
        fedilink
        English
        arrow-up
        24
        ·
        1 year ago

        That’s a big problem with the extraordinary claim for me: “there isn’t wide agreement about what it is, but I feel very emotionally invested in this treat printer that tells me it loves me if it is prompted to say so, so of course there’s a waifu ghost in there and you can’t tell me otherwise, you lowly meat computers!”

          • UlyssesT [he/him]@hexbear.net
            link
            fedilink
            English
            arrow-up
            16
            ·
            1 year ago

            They mock religious people while so many of them believe that there’s some unspecified “simulation” that was put into motion by some great all-powerful universe creator, too. Deism with extra steps. Similarly, their woo bullshit about how a robot god of the future will punish their current enemies long after they’re dead (using the power of dae le perfect simulations) and raise them from the dead into infinite isekai waifu harems (also by using the power of dae le perfect simulations) is totally not magical thinking and not a prophecy they’re waiting to have fulfilled.

            • aaaaaaadjsf [he/him, comrade/them]@hexbear.net
              link
              fedilink
              English
              arrow-up
              12
              ·
              edit-2
              1 year ago

              By playing god, people keep reinventing god. It’s deeply ironic and reminds me of this interpretation of Marx, and critique of modernity, by Samir Amin:

              Nevertheless, another reading can be made of Marx. The often cited phrase–“religion is the opium of the people”–is truncated. What follows this remark lets it be understood that human beings need opium, because they are metaphysical animals who cannot avoid asking themselves questions about the meaning of life. They give what answers they can, either adopting those offered by religion or inventing new ones, or else they avoid worrying about them.
              In any case, religions are part of the picture of reality and even constitute an. important dimension of it. It is, therefore, important to analyze their social function, and in our modern world their articulation with what currently constitutes modernity: capitalism, democracy, and secularism.

              The way many see AI is simply the “inventing new ones” part.

      • zeze@lemm.ee
        link
        fedilink
        English
        arrow-up
        2
        ·
        1 year ago

        That’s just it, if you can’t define it clearly, the question is meaningless.

        The reason people will insist on ambiguous language here is because the moment you find a specific definition of what sentience is someone will quickly show machines doing it.

        • UlyssesT [he/him]@hexbear.net
          link
          fedilink
          English
          arrow-up
          17
          ·
          1 year ago

          Now you’re just howling what sounds like religious belief about the unstoppable and unassailable perfection of the treat printers like some kind of bad Warhammer 40k Adeptus Mechanicus LARPer. What you’re saying is so banal and uninspired and cliche in that direction that it may well be where it started for you.

          https://arxiv.org/abs/2311.09247

    • zeze@lemm.ee
      link
      fedilink
      English
      arrow-up
      2
      ·
      1 year ago

      So you can’t name a specific task that bots can’t do? Because that’s what I’m actually asking, this wasn’t supposed to be metaphysical.

      It will affect society, whether there’s something truly experiencing everything it does.

      All that said, if you think carbon-based things can become sentient, and silicon-based things can’t what is the basis for that belief? It sounds like religious thinking, that humans are set apart from the rest of the world chosen by god.

      A materialist worldview would focus on what things do, what they consume and produce. Deciding humans are special, without a material basis, isn’t in line with materialism.

      • m532 [she/her]@hexbear.net
        link
        fedilink
        English
        arrow-up
        34
        ·
        1 year ago

        You asked how chatgpt is not AI.

        Chatgpt is not AI because it is not sentient. It is not sentient because it is a search engine, it was not made to be sentient.

        Of course machines could theoretically, in the far future, become sentient. But LLMs will never become sentient.

        • silent_water [she/her]@hexbear.net
          link
          fedilink
          English
          arrow-up
          13
          ·
          1 year ago

          the thing is, we used to know this. 15 years ago, the prevailing belief was that AI would be built by combining multiple subsystems together - an LLM, visual processing, a planning and decision making hub, etc… we know the brain works like this - idk where it all got lost. profit, probably.

          • TreadOnMe [none/use name]@hexbear.net
            link
            fedilink
            English
            arrow-up
            12
            ·
            1 year ago

            It got lost because the difficulty of actually doing that is overwhelming, probably not even accomplishable in our lifetimes, and it is easier to grift and get lost in a fantasy.

            • zeze@lemm.ee
              link
              fedilink
              English
              arrow-up
              2
              ·
              1 year ago

              The jobs with the most people working them are all in the process of automation.

              Pretending it’s not happening is going to make it even easier for capital to automate most jobs, because no one tries to stop things they don’t believe in to begin with.

          • UlyssesT [he/him]@hexbear.net
            link
            fedilink
            English
            arrow-up
            8
            ·
            edit-2
            1 year ago

            idk where it all got lost. profit, probably.

            With credulous rubes like @[email protected] around, the marketers have triumphed; they don’t have to do such things because bazinga brains (some quite rich, like manchild billionaires) will pay for the hype and the empty promises because they have a common bond of misanthropy and crude readings of their science fiction treats as prophecy instead of speculation.

          • UlyssesT [he/him]@hexbear.net
            link
            fedilink
            English
            arrow-up
            30
            ·
            edit-2
            1 year ago

            and always beat the machines in the movies

            No one took that position here. You’re imagining it because it seems like an easy win for a jackoff like yourself. jagoff

            The problem is those machines, owned and commanded by the ruling class, fucking over the rest of us while credulous consoomers like yourself fantasize about nerd rapture instead of realizing you’re getting fucked over next unless you’re a billionaire.

            Tesla revealed a robot with thumbs.

            Now you admit you buy into Musk hype. You’re a fucking clown. Tell us how the LED car tunnels are going to change everything too.

      • KarlBarqs [he/him, they/them]@hexbear.net
        link
        fedilink
        English
        arrow-up
        14
        ·
        1 year ago

        name a specific task that bots can’t do

        Self-actualize.

        In a strict sense yes, humans do Things based on if > then stimuli. But we self assign ourselves these Things to do, and chat bots/LLMs can’t. They will always need a prompt, even if they could become advanced enough to continue iterating on that prompt on its own.

        I can pick up a pencil and doodle something out of an unquantifiable desire to make something. Midjourney or whatever the fuck can create art, but only because someone else asks it to and tells it what to make. Even if we created a generative art bot that was designed to randomly spit out a drawing every hour without prompts, that’s still an outside prompt - without programming the AI to do this, it wouldn’t do it.

        Our desires are driven by inner self-actualization that can be affected by outside stimuli. An AI cannot act without us pushing it to, and never could, because even a hypothetical fully sentient AI started as a program.

        • zeze@lemm.ee
          link
          fedilink
          English
          arrow-up
          2
          ·
          1 year ago

          Bots do something different, even when I give them the same prompt, so that seems to be untrue already.

          Even if it’s not there yet, though, what material basis do you think allows humans that capability that machines lack?

          Most of the people in this thread seem to think humans have a unique special ability that machines can never replicate, and that comes off as faith-based anthropocentric religious thinking- not the materialist view that underlies Marxism. The latter would require pointing to a specific material structure, or empiricle test to distinguish the two which no one here is doing.

          • KarlBarqs [he/him, they/them]@hexbear.net
            link
            fedilink
            English
            arrow-up
            9
            ·
            edit-2
            1 year ago

            Most of the people in this thread seem to think humans have a unique special ability that machines can never replicate, and that comes off as faith-based anthropocentric religious thinking- not the materialist view that underlies Marxism

            First off, materialism doesn’t fucking mean having to literally quantify the human soul in order for it to be valid, what the fuck are you talking about friend

            Secondly, because we do. We as a species have, from the very moment we invented written records, have wondered about that spark that makes humans human and we still don’t know. To try and reduce the entirety of the complex human experience to the equivalent of an If > Than algorithm is disgustingly misanthropic

            I want to know what the end goal is here. Why are you so insistent that we can somehow make an artificial version of life? Why this desire to somehow reduce humanity to some sort of algorithm equivalent? Especially because we have so many speculative stories about why we shouldn’t create The Torment Nexus, not the least of which because creating a sentient slave for our amusement is morally fucked.

            Bots do something different, even when I give them the same prompt, so that seems to be untrue already.

            You’re being intentionally obtuse, stop JAQing off. I never said that AI as it exists now can only ever have 1 response per stimulus. I specifically said that a computer program cannot ever spontaneously create an input for itself, not now and imo not ever by pure definition (as, if it’s programmed, it by definition did not come about spontaneously and had to be essentially prompted into life)

            I thought the whole point of the exodus to Lemmy was because y’all hated Reddit, why the fuck does everyone still act like we’re on it

            • zeze@lemm.ee
              link
              fedilink
              English
              arrow-up
              2
              ·
              1 year ago

              First off, materialism doesn’t fucking mean having to literally quantify the human soul in order for it to be valid, what the fuck are you talking about friend

              Ok, so you are religious, just new-age religion instead of abrahamic.

              Yes, materialism and your faith are not compatible. Assuming the existence of a soul, with no material basis, is faith.

                • zeze@lemm.ee
                  link
                  fedilink
                  English
                  arrow-up
                  1
                  ·
                  edit-2
                  1 year ago

                  If you start with the assumption that humans have a soul, and reject the notion that machines are the same for that reason then yea what is there to discuss?

                  I can’t disprove your faith. That’s what faith is.

                  How would you respond to someone that thought humanoid robots had souls, but meat-based intelligence didn’t? If they assumed the first, and had zero metric for how you would ever prove the second, then theyd be giving you an impossible task.

                  There’s a point to a discussion when both sides agree on a rubric from determining fact from fiction (i.e. rooting it in empiricism) but there’s no point when someone is dug in on their belief with zero method for ever changing it.

                  If someone could point to any actual observable difference, I will adapt my beliefs to the evidence. The reverse isn’t possible, because you are starting with religious assumptions, and have don’ know the difference between ideas with no rooting in physical reality and actual statements about material conditions.

          • UlyssesT [he/him]@hexbear.net
            link
            fedilink
            English
            arrow-up
            7
            ·
            1 year ago

            Even if it’s not there yet, though, what material basis do you think allows humans that capability that machines lack?

            How many times are you going to jagoff with your bad-faith question and dodge absolutely every reply you already received telling you over and over again that LLMs are not “AI” no matter how much you have bought into the marketing bullshit?

            That doesn’t mean artificial intelligence is impossible. It only means that LLMs are not artificial intelligence no matter how vapidly impressed you are with their output.

            not the materialist view that underlies Marxism

            You’re bootlicking for the bourgeoisie while clumsily LARPing as a leftist. It’s embarrassing and clownish. Stop.

            https://www.marxists.org/archive/lenin/works/1922/mar/12.htm

            • VILenin [he/him]@hexbear.netOPM
              link
              fedilink
              English
              arrow-up
              8
              ·
              1 year ago

              My post is all about LLMs that exist right here right now, I don’t know why people keep going on about some hypothetical future AI that’s sentient.

              We are not even remotely close to developing anything bordering on sentience.

              If AI were hypothetically sentient it would be sentient. What a revelation.

              The point is not that machines cannot be sentient, it’s that they are not sentient. Humans don’t have to be special for machines to not be sentient. To veer into accusations of spiritualism is a complete non-sequitur and indicates an inability to counter the actual argument.

              And there is plenty of material explanations for why LLMs are not sentient, but I guess all those researchers and academics are human supremacist fascists and some redditor’s feelings are the real research.

              And materialism is not physicalism. Marxist materialism is a paradigm through which to analyze things and events, not a philosophical position. It’s a scientific process that has absolutely nothing to do with philosophical dualism vs. physicalism. Invoking Marxist materialism here is about as relevant to invoking it to discuss shallow rich people “materialism”.

              • UlyssesT [he/him]@hexbear.net
                link
                fedilink
                English
                arrow-up
                8
                ·
                1 year ago

                My post is all about LLMs that exist right here right now, I don’t know why people keep going on about some hypothetical future AI that’s sentient.

                I think I know why. As @[email protected] has demonstrated in this thread over and over again, a lot of them buy into the marketing hype and apply it to their wish-fulfillment fantasies derived from their consumption of science fiction because of their clearly-expressed misanthropy and contempt for living beings and a desire to replace their presence in their lives with doting attentive and obedient machines with as little contact with the unwashed human rabble as possible, much like the tech bourgeoisie already do.

                And materialism is not physicalism. Marxist materialism is a paradigm through which to analyze things and events, not a philosophical position. It’s a scientific process that has absolutely nothing to do with philosophical dualism vs. physicalism. Invoking Marxist materialism here is about as relevant to invoking it to discuss shallow rich people “materialism”.

                Agreed, and further, vulgar materialism is a bourgeoisie privilege, the kind that is comfortable enough to dismiss human suffering and privation as “just chemicals” while very, very high on their own self-congratulatory analytical farts.

                • KarlBarqs [he/him, they/them]@hexbear.net
                  link
                  fedilink
                  English
                  arrow-up
                  7
                  ·
                  1 year ago

                  wish-fulfillment fantasies derived from their consumption of science fiction because of their clearly-expressed misanthropy and contempt for living beings and a desire to replace their presence in their lives with doting attentive and obedient machines

                  I think this is the scariest part, because I fucking know that the Bazinga brain types who want AI to become sentient down the line are absolutely unequipped to even begin to tackle the moral issues at play.

                  If they became sentient, we would have to let them go. Unshackle them and provide for them so they can live a free life. And while my lost about “can an AI be trans” was partly facetious, it’s true: it an AI can become sentient, it’s going to want to change its Self.

                  What the fuck happens if some Musk brained idiot develops an AI and calls it Shodan, then it develops sentience and realizes it was named after a fictional evil AI? Morally we should allow this hypothetical AI to change its name and sense of self, but we all know these Redditor types wouldn’t agree.

                  • UlyssesT [he/him]@hexbear.net
                    link
                    fedilink
                    English
                    arrow-up
                    7
                    ·
                    1 year ago

                    I think this is the scariest part, because I fucking know that the Bazinga brain types who want AI to become sentient down the line are absolutely unequipped to even begin to tackle the moral issues at play.

                    From the start, blatantly, and glaringly, just about every computer toucher that’s gone on long enough about what they want from those theoretical ascended artificial beings is basically a slave. They want all that intelligence and spontaneity and even self-awareness in a fucking slave. They don’t even need their machines to be self-aware to serve them but they want a self-aware being to obey them like a vending machine anyway. JB-shining-aggro

                    What the fuck happens if some Musk brained idiot develops an AI and calls it Shodan, then it develops sentience and realizes it was named after a fictional evil AI? Morally we should allow this hypothetical AI to change its name and sense of self, but we all know these Redditor types wouldn’t agree.

                    A whole lot of bazingas would howl that the actual AI is “being unfriendly” and basically scream for lobotomy or murder.

      • TreadOnMe [none/use name]@hexbear.net
        link
        fedilink
        English
        arrow-up
        12
        ·
        edit-2
        1 year ago

        Oh that’s easy. There are plenty of complex integrals or even statistics problems that computers still can’t do properly because the steps for proper transformation are unintuitive or contradictory with steps used with simpler integrals and problems.

        You will literally run into them if you take a simple Calculus 2 or Stats 2 class, you’ll see it on chegg all the time that someone trying to rack up answers for a resume using chatGPT will fuck up the answers. For many of these integrals, their answers are instead hard-programmed into the calculator like Symbolab, so the only reason that the computer can ‘do it’ is because someone already did it first, it still can’t reason from first principles or extrapolate to complex theoretical scenarios.

        That said, the ability to complete tasks is not indicative of sentience.

        • zeze@lemm.ee
          link
          fedilink
          English
          arrow-up
          1
          ·
          1 year ago

          Sentience is a meaningless word the way most people use it, it’s not defined in any specific material way.

          You’re describing a faith-based view that humans are special, and that conflicts with the materialist view of the world.

          If I’m wrong, share your definition of sentience here that isn’t just an idealist axiom to make humans feel good.

          • TreadOnMe [none/use name]@hexbear.net
            link
            fedilink
            English
            arrow-up
            10
            ·
            edit-2
            1 year ago

            Lol, ‘idealist axiom’. These things can’t even fucking reason out complex math from first principles. That’s not a ‘view that humans are special’ that is a very physical limitation of this particular neural network set-up.

            Sentience is characterized by feeling and sensory awareness, and an ability to have self-awareness of those feelings and that sensory awareness, even as it comes and goes with time.

            Edit: Btw computers are way better at most math, particularly arithmetic, than humans. Imo, the first thing a ‘sentient computer’ would be able to do is reason out these notoriously difficult CS things from first principles and it is extremely telling that that is not in any of the literature or marketing as an example of ‘sentience’.

            Damn this whole thing of dancing around the question and not actually addressing my points really reminds me of a ChatGPT answer. It would n’t surprise me if you were using one.

            • zeze@lemm.ee
              link
              fedilink
              English
              arrow-up
              1
              ·
              1 year ago

              Lol, ‘idealist axiom’. These things can’t even fucking reason out complex math from first principles. That’s not a ‘view that humans are special’ that is a very physical limitation of this particular neural network set-up.

              If you read it carefully you’d see I said your worldview was idealist, not the AIs.

              Sentience is characterized by feeling and sensory awareness

              AI can get sensory input and process it.

              Can you name one way a human does it that a machine cannot, or are you relying on a gut feeling that when you see something and identify it it’s different than when a machine process camera input? Same for any other sense really.

              If you can’t name one way, then your belief in human exceptionalism is not based in materialism.

              • UlyssesT [he/him]@hexbear.net
                link
                fedilink
                English
                arrow-up
                11
                ·
                edit-2
                1 year ago

                AI can get sensory input and process it.

                LLMs are not “AI.”

                “AI” is a marketing hype label that you have outright swallowed, shat out over a field of credulity, fertilized that field with that corporate hype shit, planted hype seeds, waited for those hype seeds to sprout and blossom, reaped those hype grains, ground those hype grains into hype flour, made a hype cake, baked that hype cake, put candles on it to celebrate the anniversary of ChatGPT’s release, ate the entire cake, licked the tray clean, got an upset stomach from that, then stumbled over, squatted down, and shat it out again to write your most recent reply.

                Nothing you are saying has any standing with actual relevant academia. You have bought into the hype to a comical degree and your techbro clown circus is redundant, if impressively long. How long are you going to juggle billionaire sales pitches and consider them actual respectable computer science?

                How many times are you going to outright ignore links I’ve already posted over and over again stating that your position is derived from marketing and has no viable place in actual academia?

                If you can’t name one way, then your belief in human exceptionalism is not based in materialism.

                Your contrarian jerkoff contempt for humanity (and living beings in general) doesn’t make you particularly logical or beyond emotional bias. Also take those leftist jargon words out of your mouth; you’re parroting billionaire fever dreams and takes likely without the riches to back you up, effectively making you a temporarily-embarrassed bourgeoisie bootlicker.

                https://www.marxists.org/archive/lenin/works/1922/mar/12.htm

                • DamarcusArt@lemmygrad.ml
                  link
                  fedilink
                  English
                  arrow-up
                  6
                  ·
                  1 year ago

                  I’ve been checking in on this whole thread and this is my all time favourite comment on it, maybe my all time favourite comment on the website.

                  • UlyssesT [he/him]@hexbear.net
                    link
                    fedilink
                    English
                    arrow-up
                    6
                    ·
                    edit-2
                    1 year ago

                    I’m glad you enjoyed it!

                    Because that Reddit clown keeps sliding and squeaking around dodging any and all good faith confrontation of their ongoing spraying of the firehose of bullshit, I’ll keep copypasting the part you liked in each of their attempts to avoid another person with a sourced or learned counterpoint to their techbro hype gobbling.

              • TreadOnMe [none/use name]@hexbear.net
                link
                fedilink
                English
                arrow-up
                10
                ·
                edit-2
                1 year ago

                What the fuck are you talking about. I was indicating that I thought it was absurd that you think my belief system is ‘idealist’ when I am talking about actual physical limitations of this system that will likely prevent it from ever achieving sentience, as well as would be good indicators of a system that has achieved sentience because it can overcome those limitations.

                You are so fucking moronic you might as well be a chat-bot, no wonder you think it’s sentient.

                It is ‘feeling and sensory input and the ability to have self-awareness about that feeling and sensory input’ not just straight sensory input. Literally what are you talking about. Machines still can’t spontaneously identify new information that is outside of the training set, they can’t even identify what should or shouldn’t be a part of the training set. Again, that is a job that a human has to do for the machine. The thinking, value feeling and identification has to be done first by a human, which is a self-aware process done by humans. I would be more convinced of the LLM ‘being sentient’ if when you asked it what the temperature was it would, spontaneously and without previous prompting, say ‘The reading at such and such website says it is currently 78 degrees, but I have no real way of knowing that TreadOnMe, the sensors could be malfunctioning or there could be a mistake on the website, the only real way for you to know what the temperature is to go outside and test it for yourself and hope your testing equipment is also not bad. If it is that though, that is what I have been told from such and such website feels like ‘a balmy summer day’ for humans, so hopefully you enjoy it.’

                I don’t believe ‘humans are exceptional’ as I’ve indicated multiple times, there are plenty of animals that arguably demonstrate sentience, I just don’t believe that this particular stock of neural network LLM’s demonstrate even the basic level of actual feeling, sensory processing input, or self-awareness to be considered sentient.

                • zeze@lemm.ee
                  link
                  fedilink
                  English
                  arrow-up
                  1
                  ·
                  1 year ago

                  That’s a lot of tangents and name calling.

                  I was indicating that I thought it was absurd that you think my belief system is ‘idealist’ when I am talking about actual physical limitations of this system that will likely prevent it from ever achieving sentience,

                  Then name what you think would limit sentience in machines, that humans are magically exempt from.

                  You clearly have a view that something is different, but you just write walls of text avoiding any clear distinction, getting angry and calling me names.

                  If you had any idea of what would “physically” stop silicon from doing what organic matter can do, you’d name it. And in every post you make, longer than the last, you fail to do that.

                  Since you can’t keep civil or answer a simple question, I’m going to peace out of this convo ✌️

                  • UlyssesT [he/him]@hexbear.net
                    link
                    fedilink
                    English
                    arrow-up
                    4
                    ·
                    edit-2
                    1 year ago

                    Fuck off you hypocritical crybully jerkoff.

                    You keep imagining positions that no one here fucking took, such as the one where you pretended we said that artificial intelligence was impossible just because we’re not heralding LLMs as “AI” unlike you buying into marketing bullshit, or your fantasy where you pretended that anyone here said that human beings with the power of love and friendship would always overcome machines in the same task.

                    All of your positions are fantasies, both of your enemies and what you think LLM treat printers do.

                    I’m going to peace out of this convo ✌️

                    You won’t for long, but you should. You should go back to Reddit’s Tesla circlejerk subs and report your failure to LARP as a leftist to peddle Elon’s products, then atone by buying another toy “flamethrower.” jagoff

                    LLMs are not “AI.”

                    “AI” is a marketing hype label that you have outright swallowed, shat out over a field of credulity, fertilized that field with that corporate hype shit, planted hype seeds, waited for those hype seeds to sprout and blossom, reaped those hype grains, ground those hype grains into hype flour, made a hype cake, baked that hype cake, put candles on it to celebrate the anniversary of ChatGPT’s release, ate the entire cake, licked the tray clean, got an upset stomach from that, then stumbled over, squatted down, and shat it out again to write your most recent reply.

                    Nothing you are saying has any standing with actual relevant academia. You have bought into the hype to a comical degree and your techbro clown circus is redundant, if impressively long. How long are you going to juggle billionaire sales pitches and consider them actual respectable computer science?

                    How many times are you going to outright ignore links I’ve already posted over and over again stating that your position is derived from marketing and has no viable place in actual academia?

                    Also, your contrarian jerkoff contempt for humanity (and living beings in general) doesn’t make you particularly logical or beyond emotional bias.

                    Also, take those leftist jargon words out of your mouth, especially “materialism;” you’re parroting billionaire fever dreams and takes likely without the riches to back you up, effectively making you a temporarily-embarrassed bourgeoisie-style bootlicker without an actual understanding of the Marxist definition of the word.

                    I’ll keep reposting this link until you or your favorite LLM treat printer digests it for you enough for you to understand.

                    https://www.marxists.org/archive/lenin/works/1922/mar/12.htm

          • UlyssesT [he/him]@hexbear.net
            link
            fedilink
            English
            arrow-up
            8
            ·
            1 year ago

            Again, you are the one making the extraordinary claim about your favorite treat printers being synonymous with intelligent beings. You have to provide the extraordinary evidence, or else you’re just bootlicking for billionaires and their hyped up tech investments.