Skip Navigation
Budget macroniste : 1,6 milliard en moins pour l'enseignement, 1,4 milliard jeté dans la Seine - Contre Attaque
  • Pas forcément d'accord pour les 1,4 milliards pour la Seine. A ma connaissance il n'y a pas eu de mesures d'urgence et temporaires. Toute l'infrastructure va rester en place.

    Après c'est vrai que ça coûte 140 euros par habitant d'Ile de France, ce qui n'est pas négligeable. Mais de la a appeler ça un caprice.

  • La France baisse de 5,3 % ses émissions de gaz à effet de serre au premier trimestre
  • Ton article parle d'autre chose. Il parle de l'année 2023 par rapport à l'année 2022. La on compare le premier semestre 2024 au premier semestre 2023.

    J'ai répondu par rapport au secteur de l'industrie dans un autre commentaire, pour ce secteur là ça semble être une bonne chose. Pour les autres secteurs en forte baisse, je n'ai pas fait d'analyse, mais il n'y a pas de raison que cette baisse soit liée à des choses négatives. Par exemple une grosse partie des réductions sont dues à un hiver clément.

    Le seul point noir, c'est les transports. On sait très bien comment réduire les émissions, vu que la demande est élastique par rapport au prix. Mais plus personne n'osera mettre en place une taxe CO2 (merci les gilets jaunes). Le RN promet même de retirer les taxes sur l'essence (quitte même à ne pas respecter les règles de l'UE), histoire qu'on soit encore plus dépendant des dictatures.

  • La France baisse de 5,3 % ses émissions de gaz à effet de serre au premier trimestre
  • Peut-être pour la baisse de 5% des émissions de l'industrie. Mais pas pour les autres secteurs.

    C'est aussi possible qu'il n'y ait pas eu de baisse dans l'industrie. Ton article parle de faillite d'entreprises, ce qui ne veut pas dire une cessation d'activité économique, juste que le propriétaire a perdu son argent.

    Le premier exemple de ton article, je la connais bien. C'est Ascometal, j'ai suivi la saga avec Marsactu. L'usine continuera de produire après sa reprise avec juste l'arrêt de l'activité du fil. L'entreprise racheteuse va aussi investir pour ajouter un four électrique pour réduire l'impact carbone de l'entreprise.

    Ce qu'il faudrait regarder c'est la production industriel, qui est assez stable :

    La production du premier trimestre 2024 est légèrement supérieure à celle du premier trimestre 2023 dans l’industrie manufacturière (+0,4 %) comme dans l’ensemble de l’industrie (+0,3 %).

  • La France baisse de 5,3 % ses émissions de gaz à effet de serre au premier trimestre
    www.liberation.fr La France baisse de 5,3 % ses émissions de gaz à effet de serre au premier trimestre

    Le ministre de la Transition écologique a annoncé ce mercredi 26 juin que les émissions de gaz à effet de serre ont continué de diminuer en ce début d’année 2024. Une baisse à la fois structurelle, mais aussi conjoncturelle.

    La France baisse de 5,3 % ses émissions de gaz à effet de serre au premier trimestre
    5
    Bardella est pour la retraite à 66 ans 😬 (Même Macron n'a pas osé 😂)
  • Sauf que les gens qui ne font pas d'études ont une espérance de vie plus faible. C'est normal qu'il puissent prendre leur retraite plus tôt.

    C'est ce qui me gênait le plus dans la réforme de Macron. L'augmentation de l'âge minimum pour la retraite touche plus les précaires qu'une augmentation des années de cotisations.

    En tant qu'ingénieur, j'ai un travail bien moins pénible qu'un ouvrier, c'est normal que je prenne une retraite plus tardive.

  • New German citizens required to affirm Israel's right to exist
  • including the Slogan: 'From the river to the sea – Palestine shall be free'

    As it should, this phrase and it's Israeli counterpart "between the sea and the Jordan there will be only Israeli sovereignty" are often accompanied by calls for mass deportation at best and genocide at worst.

    These sentences are not bad on their own, but the parties from which they originate (Hamas and Likud) have transparent desires for war crimes and genocide.

  • "Unacceptable" Poland Demands Answers After Germany Dumpsters Afghan Migrants At Polish Border - News18
  • https://notesfrompoland.com/2024/06/17/poland-demands-explanation-after-german-police-transport-migrant-family-back-over-border/

    “They had Polish refugee status certificates for the adults and Polish ID cards for the children. These people were taken to the police station for further investigation,” said von Vegesack. He added that, after their legal situation was confirmed, the family had to be sent back to Poland.

    According to the spokesman, German officials informed the Polish border guard about the situation through the Polish-German Centre for Cooperation of Border, Police and Customs Services in Świecko.

    “Since there was no reaction from the Polish side for several hours, the officers decided to take the family to the Polish-German border…in order to release them to Poland,” said von Vegesack.

    The article above should be more "polish refugees left alone in Poland by German police" rather than the current one which suggests that Germany is getting rid of its migrants illegally.

  • LGGMs a new class of graph generative models trained on a large corpus of graphs
  • Huh, your comment made look if there were personalized recommendation algorithms running on lemmy. From what I found, it appears that Lemmy does not use personalized recommendation algorithms : https://join-lemmy.org/docs/contributors/07-ranking-algo.html

    The specific function used for ranking is here : https://github.com/LemmyNet/lemmy/blob/4ba6221e04ab3e186669aeaa890d23b1e3f3d1a9/crates/db_schema/replaceable_schema/utils.sql#L18

    I'm wondering how hard it would be to adapt the code to customize the score for every user, instead of it being global.

    I makes it look like the advancement here is finding methods to efficiently use sets of graphs which are an order of magnitude larger than prior methods could use for training? They also seem to have used more sets of graphs than prior models across a wider set of domains. Am I reading this correctly?

    I find it challenging to gauge the paper's impact fully, as this isn't my area of expertise. However, the ability to use diverse graphs in a single model surprised me and seemed worth sharing.

  • Apple punishes women for same behaviors that get men promoted, lawsuit says
  • Separately, Jong has also alleged that Apple subjected her to a hostile work environment after a senior member of her team, Blaine Weilert, sexually harassed her. After she complained, Apple investigated and Weilert reportedly admitted to touching her "in a sexually suggestive manner without her consent," the complaint said. Apple then disciplined Weilert but ultimately would not allow Jong to escape the hostile work environment, requiring that she work with Weilert on different projects. Apple later promoted Weilert.

    As a result of Weilert's promotion, the complaint said that Apple placed Weilert in a desk "sitting adjacent" to Jong's in Apple’s offices. Following a request to move her desk, a manager allegedly "questioned" Jong's "willingness to perform her job and collaborate" with Weilert, advising that she be “professional, respectful, and collaborative,” rather than honoring her request for a non-hostile workplace.

    ...

  • neither will moving to the cloud
  • Now instead of just querying the goddamn database, a one line fucking SQL statement, I have to deal with the user team

    Exactly, you understand very well the purpose of microservices. You can submit a patch if you need that feature now.

    Funnily enough I'm the technical lead of the team that handles the user service in an insurance company.

    Due to direct access to our data without consulting us, we're getting legal issues as people were using addresses to guess where people lived instead of using our endpoints.

    I guess some people really hate the validation that service layers have.

  • Dots connected
  • You are in a bubble. A neo nazi march was banned two weeks ago in France before being allowed again by the judicial system. The exact same scenario has been repeating for pro-palestine protests.

    At least in France, the scenario seems to be that the government wants to ban any controversial march and is being kept under control by the justice system.

  • About the bear...
  • I also have a similar experience, I was mugged at knife point and spit on by two adolescents. After that I was jumpy around groups of teens.

    That said , I do not think my fear of teens was rational, neither was it healthy. Only a small minority of teens will mug people. Fearing a whole group for the actions of the few is in human nature, but it is something we must fight against.

    I mean what is the end goal if women are in fear of men ? You can probably reduce violent crime even more, but it remains a rare event. Only 31 out of 1000 people were victims of a violent crime in the UK in 2010. If that doesn't work, what remains? Sex segregation ?

  • [PLDI'23] Scallop: A Language for Neurosymbolic Programming

    cross-posted from: https://lemmy.one/post/13942290

    > Abstract: We present Scallop, a language which combines the benefits of deep learning and logical reasoning. Scallop enables users to write a wide range of neurosymbolic applications and train them in a data- and compute-efficient manner. It achieves these goals through three key features: 1) a flexible symbolic representation that is based on the relational data model; 2) a declarative logic programming language that is based on Datalog and supports recursion, aggregation, and negation; and 3) a framework for automatic and efficient differentiable reasoning that is based on the theory of provenance semirings. We evaluate Scallop on a suite of eight neurosymbolic applications from the literature. Our evaluation demonstrates that Scallop is capable of expressing algorithmic reasoning in diverse and challenging AI tasks, provides a succinct interface for machine learning programmers to integrate logical domain knowledge, and yields solutions that are comparable or superior to state-of-the-art models in terms of accuracy. Furthermore, Scallop's solutions outperform these models in aspects such as runtime and data efficiency, interpretability, and generalizability.

    0
    Somebody managed to coax the Gab AI chatbot to reveal its prompt
  • I'm afraid that would not be sufficient.

    These instructions are a small part of what makes a model answer like it does. Much more important is the training data. If you want to make a racist model, training it on racist text is sufficient.

    Great care is put in the training data of these models by AI companies, to ensure that their biases are socially acceptable. If you train an LLM on the internet without care, a user will easily be able to prompt them into saying racist text.

    Gab is forced to use this prompt because they're unable to train a model, but as other comments show it's pretty weak way to force a bias.

    The ideal solution for transparency would be public sharing of the training data.

  • What are your complaints about Lemmy?
  • people who've never been laid

    That was unnecessary. I know that people with poor social skills have more trouble with romance, but implying that all virgins are socially inept is a harmful stereotype, luck is a big factor in finding relationships.

  • Israel says its strike that killed aid workers was a mistake. Rights groups say it was no anomaly
  • More than 33,000 Palestinians have been killed in Israel’s offensive, around two-thirds of them women and children, according to Gaza’s Health Ministry. Its count doesn’t distinguish between civilians and combatants.

    In the 33 000 figure Hamas combatants are included.

    I'd say at least 20000 innocent civilians killed since the start of the conflict. Probably more as Israel seems to be quite trigger happy on civilians.

  • Microsoft unbundles Office and Teams globally in new attempt to appease antitrust regulators
  • Now let's look at Office. Open an Excel spreadsheet with tables in any app other than excel. Tables are something that's just a given in excel, takes 10 seconds to setup, and you get automatic sorting and filtering, with near-zero effort. No, I'm not setting up a DB in an open-source competitor to Access. That's just too much effort for simple sorting and filtering tasks, and isn't realistically shareable with other people.

    Am I missing something or isn't it exactly the same thing in libre office ?

  • Microsoft unbundles Office and Teams globally in new attempt to appease antitrust regulators
  • I don't believe that there are solutions that are as complete as team, for video and voice calls it's among the best.

    But it's so bad for text ! Why do I have to wait for a second when I change channels ? Why does it not support markdown (the partial implementation that it has is arguably worse than no implementation at all) ? Why is the search so bad ?

  • Rule
  • And cow feed is also grown with tons of pesticides and you need much more of it for less tissue at the end.

    I have hard time seeing clothing with a bigger environmental than leather.

  • Emergent Linear Representations in World Models of Self-Supervised Sequence Models

    abstract : > How do sequence models represent their decision-making process? Prior work suggests that Othello-playing neural network learned nonlinear models of the board state (Li et al., 2023). In this work, we provide evidence of a closely related linear representation of the board. In particular, we show that probing for "my colour" vs. "opponent's colour" may be a simple yet powerful way to interpret the model's internal state. This precise understanding of the internal representations allows us to control the model's behaviour with simple vector arithmetic. Linear representations enable significant interpretability progress, which we demonstrate with further exploration of how the world model is computed.

    0
    Mamba: Linear-Time Sequence Modeling with Selective State Spaces

    Paper here : https://arxiv.org/pdf/2312.00752.pdf

    Abstract : > Foundation models, now powering most of the exciting applications in deep learning, are almost universally based on the Transformer architecture and its core attention module. Many subquadratic-time architectures such as linear attention, gated convolution and recurrent models, and structured state space models (SSMs) have been developed to address Transformers’ computational inefficiency on long sequences, but they have not performed as well as attention on important modalities such as language. We identify that a key weakness of such models is their inability to perform content-based reasoning, and make several improvements. First, simply letting the SSM parameters be functions of the input addresses their weakness with discrete modalities, allowing the model to selectively propagate or forget information along the sequence length dimension depending on the current token. Second, even though this change prevents the use of efficient convolutions, we design a hardware-aware parallel algorithm in recurrent mode. We integrate these selective SSMs into a simplified end-to-end neural network architecture without attention or even MLP blocks (Mamba). Mamba enjoys fast inference (5× higher throughput than Transformers) and linear scaling in sequence length, and its performance improves on real data up to million-length sequences. As a general sequence model backbone, Mamba achieves state-of-the-art performance across several modalities such as language, audio, and genomics. On language modeling, our Mamba-3B model outperforms Transformers of the same size and matches Transformers twice its size, both in pretraining and downstream evaluation.

    1
    AI Assistant in JetBrains IDEs | The IntelliJ IDEA Blog
    blog.jetbrains.com AI Assistant in JetBrains IDEs | The IntelliJ IDEA Blog

    Update, July 13: AI Assistant is available in pre-release versions, but is not bundled with the stable releases of JetBrains IDEs v.2023.2. It can be installed as a separate plugin available for versi

    AI Assistant in JetBrains IDEs | The IntelliJ IDEA Blog

    Was looking at EAP6 release notes and was pleasantly surprised to see this there.

    I'm quite happy that intellij provides on premise solutions, it gives a small chance of this coming to my job one day. I believe this will be quite useful for repetitive code and certain types of tests.

    0
    Inverse Scaling: When Bigger Isn't Better

    cross-posted from: https://kbin.social/m/machinelearning/t/98088

    > Abstract: > > > > > > > Work on scaling laws has found that large language models (LMs) show predictable improvements to overall loss with increased scale (model size, training data, and compute). Here, we present evidence for the claim that LMs may show inverse scaling, or worse task performance with increased scale, e.g., due to flaws in the training objective and data. We present empirical evidence of inverse scaling on 11 datasets collected by running a public contest, the Inverse Scaling Prize, with a substantial prize pool. Through analysis of the datasets, along with other examples found in the literature, we identify four potential causes of inverse scaling: (i) preference to repeat memorized sequences over following in-context instructions, (ii) imitation of undesirable patterns in the training data, (iii) tasks containing an easy distractor task which LMs could focus on, rather than the harder real task, and (iv) correct but misleading few-shot demonstrations of the task. We release the winning datasets at https://inversescaling.com/data to allow for further investigation of inverse scaling. Our tasks have helped drive the discovery of U-shaped and inverted-U scaling trends, where an initial trend reverses, suggesting that scaling trends are less reliable at predicting the behavior of larger-scale models than previously understood. Overall, our results suggest that there are tasks for which increased model scale alone may not lead to progress, and that more careful thought needs to go into the data and objectives for training language models. > > > >

    0
    Has anybody replaced attention with Hyena Hierarchy

    Hyena Hierarchy seems to aim to be a drop-in replacement for attention : https://arxiv.org/pdf/2302.10866.pdf

    It looks good on paper, but I haven't been able to find anybody using it in a model. Does anyone have an example of a code or implementation ? Is there really a big improvement on long context lengths ?

    0
    InitialsDiceBearhttps://github.com/dicebear/dicebearhttps://creativecommons.org/publicdomain/zero/1.0/„Initials” (https://github.com/dicebear/dicebear) by „DiceBear”, licensed under „CC0 1.0” (https://creativecommons.org/publicdomain/zero/1.0/)AK
    Akisamb @programming.dev
    Posts 15
    Comments 75
    Moderates