Category: Universities

  • MIL-OSI Russia: Fusion of academic and practical: students and postgraduates explore creative economy

    Translation. Region: Russian Federation –

    Source: State University Higher School of Economics – State University Higher School of Economics –

    The HSE hosted The Fourth International Forum of Young Researchers in Creative Economy. The authors of scientific papers that passed the competitive selection presented their reports: Russian and foreign researchers under 35 years of age, postgraduates and students studying the socio-economic aspects of the development of the creative economy and creative industries.

    The Fourth International Forum of Young Researchers in the Creative Economy took place on the second day of the IV International Scientific Conference “Creative Economy: Key Development Trends and State Policy”. The event is being held Institute for Statistical Research and Economics of Knowledge (ISSEZ) under the auspices of the Decade of Science and Technology in Russia, within the framework of the activities of the World-Class Scientific Center “Center for Interdisciplinary Research on Human Potential» HSE University, with the support of the Russian Ministry of Education and Science.

    The researchers studied key trends in the development of the creative economy and creative industries, digitalization of creative industries, analysis of creative clusters, creative potential of cities and regions, etc.

    Every year more and more people participate in the competition, and now the competition was serious, emphasized the forum moderator, director of the center “Russian Cluster Observatory» Institute for Statistical Studies and Economics of Knowledge, National Research University Higher School of Economics Evgeny Kutsenko.

    The three best works were awarded diplomas and prizes. Elizaveta Fainshtein from the National Research University Higher School of Economics studied the perception of visitors to various public spaces based on a semantic analysis of reviews. The jury noted the relevance of the work and its practical focus. “Such spaces are appearing in many cities in Russia, and your research shows how they can help themselves, because, of course, there is not always enough money to develop them through the state,” noted Evgeny Kutsenko.

    The work of Ivan Slipchenko from the Central University of Finance and Economics (China) was devoted to the impact of government support measures on the dynamics of China’s creative goods exports in 2016–2023. The researcher analyzed which support measures are most important and used a panel model for this purpose, which allows assessing cause-and-effect relationships.

    Timur Malikov from the National Research University Higher School of Economics studied the reasons for overtime work in the video game industry. The jury noted the deep study of the topic, as well as the high-quality fusion of academicism and practicality.

    In addition, the members of the competition committee additionally singled out four more works that they liked. A team of HSE students presented the study “Bread and Wine: Defining the Boundaries of Influence of Modern Creative Clusters in Moscow Using the Example of Spaces Near the Dmitrovskaya and Kurskaya Metro Stations”. Another team of HSE students studied the phenomenon of “catch-up” creativity in northern regions using the example of the Murmansk Region. Vitaly Saakov from the Russian State University of Economics (RINH) conducted an analysis of the creative industries of the Rostov Region. Anastasia Makukhina from the State Institute of Art Studies studied social networks as a factor in shaping demand for theatrical goods.

    In conclusion, Evgeny Kutsenko called on the contestants to refine their research in accordance with the recommendations and take part in the HSE competition next year.

    Please note: This information is raw content directly from the source of the information. It is exactly what the source states and does not reflect the position of MIL-OSI or its clients.

    MIL OSI Russia News

  • MIL-OSI USA: Press Release: FDIC Appoints Amanda J. Lavis as Director of Office of Equal Employment Opportunity

    Source: US Federal Deposit Insurance Corporation FDIC

    WASHINGTON – The Federal Deposit Insurance Corporation (FDIC) today announced its Board of Directors has approved the appointment of Amanda J. Lavis as Director of the agency’s new Office of Equal Employment Opportunity (OEEO).

    In June, the FDIC Board announced the creation of the OEEO to serve as a single point of entry for employee complaints of discrimination and retaliation.  In this role, Ms. Lavis will lead the OEEO’s work to intake, investigate, and report on complaints of employment discrimination within the FDIC workplace.  The OEEO, along with the agency’s new Office of Professional Conduct (OPC), will report directly to the FDIC Board.

    Ms.  Lavis was selected from among several highly qualified candidates after a competitive, nationwide public solicitation.  Most recently, she served as Chief Culture Officer for the U.S. Army Combat Capabilities Development Command (DEVCOM), where she was the primary EEO advisor to the Commanding General and executive leadership.

    She previously served as the State of Hawaii’s EEO Officer, serving as the primary EEO advisor to the Governor and Executive Branch.  As an attorney and partner at Rhoads & Sinon LLP, Ms. Lavis worked with financial institutions and other private and public sector clients on employment issues and all aspects of EEO compliance.

    Ms. Lavis has a Juris Doctor from the Villanova University School of Law, a Master of Business Administration from Shippensburg University, and a Bachelor of Science in International Business from Messiah University.

    # # #

    MEDIA CONTACT: 
    mediarequests@fdic.gov


    FDIC: PR-94-2024

    MIL OSI USA News

  • MIL-OSI Economics: Advancing prosperity in the age of AI

    Source: Microsoft

    Headline: Advancing prosperity in the age of AI

    As we approach another national election in the United States, both the country and the world are rightly focused on what comes next. The next president of the United States, along with new leaders in countries like the United Kingdom and Japan, will need to navigate economic and climate challenges, societal divides, and international conflicts. Looking more broadly, the next four yearsand indeed the next quarter-centurywill be marked by rapid technological change. This means that success for nations and the world will depend on our collective ability to manage this change well. 

    Today, we are at the threshold of major advances in life sciences, energy, and climate technology. However, the most significant opportunities in the second quarter of the 21st century will almost certainly be driven by advances in artificial intelligence (AI). This underscores the imperative for countries to develop national strategies and policies that effectively harness AI’s potential. For these strategies to succeed, it’s essential that we recognize AI’s role as a general-purpose technology and promote investments that support its broad adoption across the economy, including skilling initiatives that will position citizens to thrive in the new age of AI. 

    The World’s Next Great General-Purpose Technology 

    Economists categorize technologies into two types: single-purpose tools and general-purpose technologies, or GPTs. A single-purpose tool, like a smoke detector or lawn mower, excels at one specific task. But general-purpose technologies, like electricity or personal computers, have multiple applications and can be utilized across every economic sector. As we look ahead, it’s almost certain that AI will be regarded by economists as the next great GPT. 

    GPTs are transformative. They have the power to reshape economies and societies. A new book by Jeffrey Ding, a professor at George Washington University, documents the extraordinary degree to which GPTs have reshaped economies and even the economic balance among nations.  

    In “Technology and the Rise of Great Powers”, Professor Ding reviews the impact of GPTs over the past 250 years. He documents how the First Industrial Revolution, beginning in the United Kingdom in the 18th century, was defined by mechanization of agriculture and manufacturing based on ironworking, the most impactful GPT of the time. The Second Industrial Revolution, in the late 19th century, catapulted economic growth in the United States through the widespread adoption of two new GPTs: electricity and machine tools. The Third Industrial Revolution, which began in the 20th century, was driven by a new generation of GPTs—computerization and digital technologies—with the United States again leading the world in technology adoption. 

    Perhaps most importantly, Professor Ding documents a phenomenon that may surprise some policymakers but is familiar to many in the tech sector. He explains that the most important long-term determinant of a country’s economic growth during an industrial revolution is not whether it is at the forefront of innovation in a “leading sector” of the time. Instead, it’s whether the country “diffuses”—or spreads—the adoption of a critical GPT broadly across its economy.   

    This conclusion is intuitive, given that historically critical GPTs significantly boost productivity. The more widely a GPT is adopted, the greater its contribution to the productivity gains that drive economic growth. While it’s possible for a nation to have an advantage in both leading sector innovation and broad GPT adoption, Microsoft’s first-hand experience suggests that the sustained economic growth of nations in the first quarter of the 21st century is most closely linked to the widespread and consistent adoption of digital technologies. 

    This insight has profound implications for the impact of AI over the next 25 years. Today, policymakers in some capitals—and especially Washington, D.C.—are focused almost single-mindedly on whether their country can control and dominate cutting-edge innovation in new leading sector technologies such as graphical processing units and frontier AI models. While these are important policy issues, it’s equally, if not more, important to address what it will take to ensure the widespread and effective adoption of AI across all the societal sectors that can benefit from it. 

    Another important insight from the impact of GPTs over time is the contrast between early innovation and the delay in widespread technology adoption. The early stages of innovation often feel like an intense and even short-lived race to the technology visionaries involved, whether they are the inventors of electricity, automobiles, computers, or AI. However, broad technology adoption takes more time. Even innovations that advanced the cutting edge of technology in years required broad societal adoption that took decades. There are many reasons to believe that this pattern will hold true for AI. 

    That’s why it’s crucial to look forward now, both at the remainder of this decade and at the upcoming second quarter of the century. Countries will need to combine short and long-term strategies to be successful. These strategies will require multiple components, two of which I discuss here. 

    Building AI Skills 

    One of the vital lessons from history is the role of skilling in spreading the adoption of a critical GPT. Organizations across an economy cannot adopt new technology unless they have the skilled workers needed to use it. 

    I witnessed this firsthand during the early expansion of the PC sector. Before joining Microsoft in 1993, I spent four years in London as a lawyer helping the American PC software sector expand across Europe. In each country, this initial growth required two key components: the protection of software under copyright law to ensure organizations paid for it and investment in skilling programs to equip people with the skills to use it. 

    It’s easy to forget today that the early years of personal computing required users to study manuals or attend a class to learn how to use a computer or a new software application. When I bought my first computer in 1985, I kept a small library of manuals next to my PC, including Microsoft Word 1.0. Employers worldwide invested in PC training for their employees, but no country embraced this more broadly and rapidly than the United States between 1980 and the year 2000. 

    I recalled this experience when two weeks ago we brought more than 2,000 Microsoft employees from around the world to Seattle for a week of meetings that kicked off with a day of professional development classes. These included six different courses for non-technical employees on how to get the most from our Copilots and other AI applications. These classes were designed to help us bridge the gap between our current abilities and the evolving needs of the AI-driven workplace. While we live in a world with broad digital fluency and a vital computer science profession, the age of AI will require new efforts to learn the latest AI skills.  

    Professor Ding’s book illustrates that the need for new skills has been critical to the spread of all major GPTs since the 1700s. This extends well beyond the needs of everyday users, highlighting that an advanced skilling infrastructure is indispensable in expanding the professions that create applications that make broad use of new technologies. 

    For example, ironworking in the 1700s spread more rapidly in the United Kingdom than elsewhere because technical associations and apprenticeships in the country enabled workers to master new skills. Machine tooling in the late 1800s spread more quickly in the United States because land-grant colleges expanded the number of mechanical engineers. And the adoption of digital technology in the U.S. over the past 50 years has also benefited enormously from the rapid growth of computer science departments across American college campuses. 

    The second quarter of the 21st century will require countries to develop national AI skilling strategies. These strategies must build upon existing disciplines like computer and data science, projecting how these fields will evolve into jobs and careers for AI engineers and AI systems designers, among others. They also will need to reflect the broader array of AI fluency across different parts of the economy. And national strategies will need to build on existing educational infrastructure and determine the best ways to provide skilling opportunities across various economic sectors. 

    The Role of Social Acceptance 

    Another historical lesson involves the critical role of social acceptance of technology. This too reflects common sense: new technology never becomes truly important unless people want to use it.  

    Academic research in the 20th century made significant strides in understanding why some technologies spread more rapidly than others. Public or social acceptance typically comes down to two factors: usefulness and trust. Technologies must solve real-world problems and improve people’s lives. At the same time, they must be trustworthy, with safeguards in place to protect a country’s societal and ethical values. 

    When put in this light, it’s easy to understand why the early years of electricity involved such intense competition between Thomas Edison, George Westinghouse, and Nikola Tesla over the safety implications of different types of electrical currents. Each inventor was trying to prove that its approach was the safest and most reliable. They knew people would only use technology they trusted.  

    This provides important context for the evolution of both industry practices and government regulation of AI. The widespread adoption of AI will in part turn on the continued development of corporate governance models to ensure that AI is used safely, securely, and in a manner that the public regards as trustworthy. Companies that develop and deploy AI must continue to invest in AI governance processes and practices that earn the public’s trust.  

    While government leaders will change over time, every nation must continue to pursue balanced efforts to develop laws and regulations that govern these aspects of AI. Sustained public trust depends on it. And the ability for countries around the world to adopt AI broadly and inexpensively will require regulatory interoperability and consistency to ensure that AI advances in one country can move to other like-minded nations. 

    Broad social acceptance for AI will likely depend on three more factors. First, we need to ensure that AI creates new opportunities for workers, not just productivity growth. While this starts with broad AI skilling, it cannot stop there. Technology adoption across an organization requires thoughtful change management, and the most effective approaches typically involve input from the workers who will put it to work. There is a lot of room for new and innovative partnerships to spread best practices in this area, both among employer associations and with organized labor. 

    Second, the tech sector needs to take a responsible approach to AI competition issues. Elected and appointed officials will change, but if we look forward with the time horizon of the quarter century ahead, it’s apparent that governmental questions and proceedings will remain a fact of life—as they have since the United States adopted the Sherman Act to govern antitrust law in 1890 in reaction to the Second Industrial Revolution. Ultimately, public confidence in new technology requires confidence in the market that creates it. 

    This perspective is part of what led Microsoft to draft and adopt 11 AI Access Principles in February. These voluntary principles are designed to ensure open access, fairness, and responsibility as we deploy AI infrastructure, platforms, and applications around the world. We’re obviously not alone in thinking about these issues, and as always, governments will play the determinative role. This past year alone, the UK’s Competition and Markets Authority (CMA) adopted cutting-edge AI Principles, and the European Commission continues to focus on the application of its Digital Markets Act to AI. Plainly, these will represent an important part of the developments ahead. 

    Finally, social acceptance of AI will likely require a consistent focus on the impact of AI on another paramount challenge of our era: climate sustainability. We are optimistic about the ways that AI can help pursue new advances in climate technology and practices. However, we are also keenly aware that AI requires the construction of more datacenters and the use of more electricity. Both as companies and in partnership with governments, we need to conserve water and reduce carbon emissions. That’s why we’re investing as a company in greener technologies such as carbon-free sources of electricity and eco-friendly steel, concrete, and fuels. 

    The Path Forward 

    Ultimately, the world needs AI that is not only more powerful but also broadly accessible and trustworthy. Between now and the midpoint of the 21st century, countries can harness AI to enhance both productivity and prosperity.  

    We shouldn’t be pollyannish. Challenges are inevitable, as history shows. New leaders, both now and in the decades ahead, will need to navigate these challenges with thoughtfulness and agility. 

    But the opportunities ahead are far greater than the challenges. We can learn from history to ensure that AI creates benefits that are shared widely. Countries can invest in the skilling infrastructure needed for success. And across the public and private sectors, we can work together to earn and sustain public acceptance for the next great GPT that will not just shape but define a critical aspect of the quarter century ahead. 

    Tags: Accessibility, AI, AI for Accessibility, AI for Good, Governance, Responsible AI

    MIL OSI Economics

  • MIL-OSI NGOs: UK: ‘Apartheid, Occupation, Genocide’ panel event with leading international law experts on Israel/Palestine

    Source: Amnesty International –

    Amnesty International UK and the International Centre of Justice for Palestinians are hosting a panel event in central London on Tuesday 12 November with four leading international law experts to discuss the worsening human rights crisis in Israel and the Occupied Palestinian Territory.

    The expert panel – comprising Zaha Hassan, Gerhard Kemp, Itay Epshtain and Victor Kattan – will discuss how enforcing international law can help secure justice for Palestinians.

    The panellists will be available for media interviews both before and after the event.

    The event is being held at 1 Birdcage Walk in central London. Further information and tickets are available here

    Event details

    What: Amnesty International UK and the International Centre of Justice for Palestinians panel discussion with the following speakers:

    Zaha Hassan, Palestinian human rights lawyer, and former coordinator and senior legal advisor to the Palestinian negotiating team during Palestine’s bid for UN membership

    Gerhard Kemp, Professor of law at UWE Bristol Law School, and extraordinary professor of public law at Stellenbosch University, South Africa

    Itay Epshtain, Special Advisor on International Law and Humanitarian Principles to the Norwegian Refugee Council, and former director of Amnesty International Israel

    Victor Kattan, Assistant Professor of Public International Law at the University of Nottingham School of Law

    Where: 1 Birdcage Walk, Westminster, London SW1H 9JJ

    When: Tuesday 12 November 2024, 19:00-21:30

    View latest press releases

    MIL OSI NGO

  • MIL-OSI Global: Why the chancellor’s plan to unlock billions of pounds of government investment is such a gamble

    Source: The Conversation – UK – By Steve Schifferes, Honorary Research Fellow, City Political Economy Research Centre, City St George’s, University of London

    Perhaps the most important long-term change announced in the first Labour budget are the new rules the government has set itself to fund the expansion of public services and increase public investment. These fiscal rules, which set out how much the government can borrow and spend, are seen as critical to reassuring the markets and the public that the government is sensibly managing the economy.

    Labour has long claimed that former prime minister Liz Truss casting aside the rules to introduce unfunded tax cuts in 2022 wrecked the British economy and left families worse off with higher mortgage and borrowing costs. Chancellor Rachel Reeves came into office determined to show that Labour would be fiscally responsible.

    The government says this budget will make working families better
    off. In its own analysis, it shows that only the top 10% of the income distribution are made worse
    off (by 1%) by the plans. The poorest households gain the most (by 5%). However, this analysis counts benefits from the big increase in public spending on areas like health and education, which tend to be used more (relative to their income) by poorer households.

    Actual cash income offers a different picture. Spending watchdog the Office for Budget Responsibility (OBR) argues that 75% of the change to employers’ national insurance will be passed on to workers in lower wages (although the minimum wage will be boosted by 6.7% to £12.21 an hour). And there is very little for the working poor or those outside the labour market on universal credit (although pensioners have been protected).

    This budget was delivered against the background of two big challenges that need urgent action: the parlous state of the public sector after years of austerity, and the very slow growth of the UK economy, which has meant little increase in real incomes.

    To deal with these two issues, Reeves made some big changes to the previous government’s fiscal rules. This will give her space to borrow more money to finance public investment – spending on things like roads, hospitals and emerging industries that should feed into economic growth.

    Finding the money

    She has done this firstly by changing the so-called “fiscal mandate”, which relates to how much the government can borrow in any individual year. Under the new rule, within three years the government must get as much back in taxes as it spends (excluding investment).

    It is the need to meet this rule that means the government has to raise taxes by £40 billion (more than half from the increase in employers’ national insurance contributions) to fund the spending needed to run the NHS, education and other public services.

    But the government has another rule to prevent the total amount of government debt becoming too large compared to the size of the economy as a whole (GDP). Here the chancellor has chosen to change how government debt is defined, adding some more government financial assets, such as money put aside for local government pensions and student loans, to set against the outstanding amount being borrowed.

    This has given her the room to borrow an extra £50 billion a year for investment, although she plans to use only half of that. The hope is that more public investment will both boost the economy (for example, by providing more roads and green energy) and improve public sector productivity (by providing things like more schools, health centres and scanners).

    Investment in equipment would lead to increased productivity within the NHS.
    l i g h t p o e t/Shutterstock

    The OBR has judged that Reeves will meet her self-imposed rules within three years, despite the huge £70 billion increase in government spending. But it warns that the margin for error is quite small for both measures. The OBR also suggests that the economic benefits of increased public investment could take a long time to materialise, well beyond the five-year forecast period.

    There are other risks to Reeves’ strategy. The cost of borrowing could go up if those financial institutions that lend the government money demand a higher interest rate.

    The OBR projects that the government will be spending £100 billion a year on debt interest payments for each of the next five years. While the large increase in government spending and borrowing will initially boost the economy, it also means inflation is likely to stay slightly higher as more money is pumped into the economy. This, of course, could slow the rate at which the Bank of England cuts interest rates.

    Gains for the population as a whole over the five-year parliament appear to be modest, with the second smallest rise in household income of any recent parliament of just 0.5%. This is driven by OBR projections that the budget will not initially boost growth very much despite greater borrowing.

    And if the economy does not grow as much as hoped, the government may need more money to meet its day-to-day costs – especially as much of the new money has been front-loaded to be spent in the next two years. This would necessarily increase taxes even further.

    The fiscal rules mirror Labour’s political dilemma, the need for short-term pain in order to get long-term gains in improved public services, a more productive economy and higher incomes and living standards. What is not clear is how long the public will wait to see results.

    If, by the end of the parliament, people don’t feel like they have more in their pockets despite all the additional spending then Labour’s credibility could be in jeopardy.

    Steve Schifferes does not work for, consult, own shares in or receive funding from any company or organisation that would benefit from this article, and has disclosed no relevant affiliations beyond their academic appointment.

    ref. Why the chancellor’s plan to unlock billions of pounds of government investment is such a gamble – https://theconversation.com/why-the-chancellors-plan-to-unlock-billions-of-pounds-of-government-investment-is-such-a-gamble-242556

    MIL OSI – Global Reports

  • MIL-OSI Global: Four ways Mohamed Al Fayed silenced whistleblowers in his organisation

    Source: The Conversation – UK – By Kate Kenny, Professor of Business and Society, University of Galway

    Mohamed Al Fayed owned the luxury goods department store Harrods from 1985 to 2010. Fred Duval/Shutterstock

    On the first anniversary of former Harrods owner Mohamed Al Fayed’s death, more than 20 women accused the billionaire of rape, sexual assault or harassment while they worked at his luxury department store. Many had been in their late teens and early twenties at the time.

    Since then, a further 65 women have come forward to the BBC with allegations dating back as far as 1977, and 40 people are reported to have contacted the police.

    How did Al Fayed silence potential whistleblowers for such a long time? I’ve researched whistleblowing in organisations for almost 15 years. Looking at the allegations made against him, four apparent strategies stand out as textbook examples of how leaders can suppress dissent to continue their terrible behaviour – even today.

    1. The organisation as a fortress

    As the chairman-owner of Harrods, Al Fayed could wander around its swanky shopping halls and oak-panelled offices as he pleased. And it appears he looked for women to target as he did so.

    Security guards had their role, in some cases reportedly turning a blind eye to distraught and dishevelled women leaving Al Fayed’s apartments and houses after attacks. HR people might likewise focus on recruiting certain women – like the security staff, they were just getting on with their work.

    That is the thing about bureaucracies, as philosophers from Hannah Arendt to Max Weber have highlighted. Staff are not responsible for the outcome. They just need to do their job.

    My research on whistleblowing in financial services shows clearly that the kind of blind rule-following many organisational roles require stops workers questioning the big picture and acting ethically by stepping in.

    2. Hi-tech surveillance

    The IRA bomb that exploded in Harrods’ car park in 1983 led to a top-notch system of surveillance being installed by its then owners.

    So, when Al Fayed bought the store two years later, his need for control was satisfied with cameras and recording systems. Eventually, everyone working at Harrods apparently knew about the system, which appears to have stopped them talking to each other about Al Fayed’s behaviour.

    Shockingly, the former Harrods owner appears to have extended this surveillance to the very bodies of the women he targeted. Doctors associated with the company were said to administer mandatory gynaecological examinations to female staff. Fayed was reportedly sent their test results. This meant he had eyes on his workers, bodies and all.

    Today, with things like social media and the ability to share large amounts of data rapidly, it is more difficult for organisations to keep information in-house. And so, we have seen a rapid growth in insider threat detection – using technology like keystroke monitoring, where every keystroke on a computer is tracked without the user’s knowledge, to identify potential leaks.

    A byproduct has been a “chill effect” on workers speaking out about wrongdoing they see in their organisations – something that has been highlighted by the UN as a problem for society.

    My research alongside other academics into whistleblowing in healthcare, engineering and government shows one thing clearly: if trust in the organisation is lacking and workers do not feel protected against potential reprisals, they stay silent. Overt surveillance deters disclosures of organisational abuses.

    Al Fayed was said to prowl Harrods on the hunt for women to target.
    DaLiu/Shutterstock

    3. Legal pressure

    The “non-disclosure agreement plus settlement payoff” tactic that Al Fayed employed with a number of Harrods staff was straight out of the Harvey Weinstein playbook. The disgraced film producer used non-disclosure agreements systematically to silence survivors.

    While non-disclosure agreements are not allowed to be used to stop workers reporting possible crimes or serious wrongdoings, a frightened 20-year-old is not likely to know this.

    In the case of Al Fayed, when Vanity Fair magazine published victims’ testimonies and allegations of serious criminality, his lawyers knew the solution. Keep the legal pressure on until the magazine settled.

    The use of legal tools to silence whistleblowers is one of the biggest concerns for researchers today. From “Slapp” suits – strategic lawsuits against public participation, filed against people who speak out – to inappropriate use of non-disclosure agreements, defensive organisations increasingly turn to the law in public whistleblowing cases. As analysis of the case of whistleblowers at the disgraced blood testing firm Theranos made clear, often the threat of legal action is enough to keep a worker silent.

    4. Dehumanise targets

    Al Fayed, we are told, would chuckle as he openly groped women. One woman reported his laughter after an attempted rape at his Villa Windsor in Paris, when he fell on the floor after she pushed him off.

    Most people would not find humour in such situations, unless they don’t see their victims as “real people”.

    But the likelihood of targets speaking out is, again, slim. A very young person told they are worthless, treated as such, and reminded of it regularly by colleagues and bosses, is not best placed to speak up. Our research with other survivors in work organisations shows how the experience of sexual violence and harassment can leave them vulnerable. They find disclosure of the abuse intolerable without empathetic and supportive colleagues.

    In an organisation designed to prevent workers discussing their concerns together – as Harrods appears to have been – the solidarity required to speak out and be protected through the collective is utterly absent.

    Harrods’ current owners have said they are “appalled” at the allegations, and the business has reached settlements with many of the people who have complained.

    When executing a campaign of “attack, isolate and silence”, money and influence can buy predators a lot of leeway, as other high-profile abusers like Weinstein and Jimmy Savile figured out. But the key thing is the organisation. With the right PR, surveillance, HR and lawyers to take legal action should stories get published, predators will be safe. The secret stays kept – until, one day, people have finally had enough.

    Kate Kenny does not work for, consult, own shares in or receive funding from any company or organisation that would benefit from this article, and has disclosed no relevant affiliations beyond their academic appointment.

    ref. Four ways Mohamed Al Fayed silenced whistleblowers in his organisation – https://theconversation.com/four-ways-mohamed-al-fayed-silenced-whistleblowers-in-his-organisation-240936

    MIL OSI – Global Reports

  • MIL-OSI Global: Recruiting the world’s first disabled astronaut doesn’t mean space travel is inclusive – here’s how to change that

    Source: The Conversation – UK – By Sean Cullen, Lecturer in Engineering Manufacturing, College of Engineering, Design and Physical Sciences, Brunel University of London

    In the past, spaceflight was the preserve of government-funded astronauts who had to meet stringent physical, cognitive, psychological and social requirements for selection. But in recent years, that has all been changing.

    In September 2024, two non-professional astronauts completed the first privately funded spacewalk, using the Crew Dragon spacecraft built by Elon Musk’s company SpaceX. Meanwhile, Houston-based private company Axiom Space is conducting regular flights to the International Space Station (ISS), carrying a mixture of government-funded astronauts and paying customers.

    In the last few years, nearly 100 people have become private astronauts through the space tourism companies Blue Origin, operated by Jeff Bezos and Virgin Galactic, by Sir Richard Branson. While the price of a seat on these vehicles remains out of reach for most of us, prices are expected to drop as more players enter the market.

    Despite the rapid growth in the number of space travellers, underrepresented population groups are still left behind, particularly those with disabilities. So how can space agencies and “space tourism” companies make spaceflight more inclusive for disabled astronauts?

    The European Space Agency (Esa) recently recruited John McFall, who lost his right leg aged 19, as the world’s first disabled astronaut. McFall, who is a surgeon and former paralympic sprinter, will participate in a feasibility study to improve understanding of, and overcome, the barriers that spaceflight presents for astronauts with physical disabilities.

    Esa’s most recent selection of astronauts was entirely of white European background, showing how far things still have to go. But its move to recruit McFall marked a significant milestone towards a more inclusive approach to spaceflight.

    Designing effective systems for the inclusion of disabled people is a longstanding challenge on Earth – and space presents a whole new paradigm. The very specific demands of spaceflight mean we can’t assume that traditional adjustments and assistive technology will work beyond Earth’s atmosphere. So, making spaceflight more inclusive requires looking at each step of going into space.

    Astronaut training is a complex process, designed to simulate the space environment and enable candidates to perform well under a variety of conditions they may encounter in orbit. But in many cases, the training facilities are not well designed for individuals with physical or sensory impairments.

    For example, in order to get on the plane that flies in an arc to simulate microgravity (colloquially referred to as the “vomit comet”), astronauts must climb a set of stairs, which presents a hurdle to anyone with a mobility impairment. Ironically, impairments that restrict the use of stairs on Earth might be much less of a restriction once in space.

    Spacecraft and space suit design will be another key focus. The space suits onboard the ISS were originally designed with male astronauts in mind, meaning that female astronauts have to “make do” with what is there. This has caused challenges as the number of female astronauts has risen.

    Older spacesuits were designed with male astronauts in mind.
    Nasa / Mike Hopkins

    In 2019, Nasa had to postpone the first all-female spacewalk because the torso of a space suit was too large for one of the spacewalkers. The Moon suit developed by Axiom Space in collaboration with Italian fashion house Prada is a step towards inclusivity, with anthropomorphic sizing to accommodate a wide range of crew members. Yet, future disabled astronauts might still encounter challenges if they have differences in their limbs or impairments to their dexterity.

    Interestingly, the new SpaceX Extra Vehicular Activity (EVA) suits have something called “embedded modularity” – each section of the suit is customised to the intended astronaut, and all sections fit together. While intended to help with joint positioning, these suits present a unique opportunity to support disabled astronauts with limb differences.

    Inclusive suits could include a single fixed leg portion for individuals with paralysis, and removable parts for those with limb differences. Haptic gloves could provide tactile feedback through the space suit for astronauts with limb differences.

    For individuals with visual impairments, incorporating augmented reality (AR) heads-up displays (transparent displays that show the user data overlaid over their environment) and AI-powered image-to-voice software that can translate purely visual information into audio explanations could make a huge difference.

    Technological support similar to the app “Be My Eyes”, pairing sighted assistants with visually impaired people to help explain their environment, could also find uses in spacesuits.

    Exercise equipment need adjustments to allow them to be used by disabled astronauts.
    NASA

    Thriving in space

    An often overlooked part of astronaut life is maintaining physical fitness through intensive exercise regimes. Exercise is required because both muscle and bone waste away quickly in microgravity – but the fitness equipment aboard the ISS, such as the treadmill and bike, is difficult to adapt for disabled people. Both require use of both feet to operate.

    Re-engineering the systems for exercise, eating, working, going to the toilet and other essential activities is critical for enabling disabled astronauts to thrive in space.

    Assistive technologies that could be used inside a spacecraft, as opposed to within a spacesuit, are continually evolving and taking many forms. As such, there are always opportunities to improve the environment on a space mission to make it more inclusive for disabled astronauts.

    Examples could include virtual reality (VR) for use in ground training, smart prosthetics that enable the completion of complex tasks, and computer vision with AI guiding visually impaired astronauts.

    Policies implemented by space agencies have traditionally been exclusionary, focusing on able-bodied individuals and ignoring the potential of those who are different. And while some space agencies are establishing advisory committees and promoting diversity, this work is often limited to narrow purposes within these agencies.

    Despite the UK and many other countries having specific laws to reduce discrimination in the workplace, the international nature of the space sector can cause difficulty. For this reason, policies mandating inclusion and equity across the space sector are crucial. Most importantly, space agencies should ensure adequate funding and resources to support any inclusion initiatives and work with disability advocacy groups.

    Often, the root causes of inclusion barriers are a lack of understanding or awareness of disabilities. In many cases, consulting and involving disabled people in decision-making processes reduces these barriers. It is essential the space sector recruits individuals from diverse backgrounds to begin with.

    Although the concept of “diversity quotas” has historically been divisive, the first place to start is to understand the diversity both of current and potential space travellers. Publicising diversity statistics can help hold agencies accountable, and encourage initiatives aimed at greater inclusion.

    There remains a lot to do, but with a collaborative approach, the new commercial space race could act as a shining example to the rest of the world in its approach to disability.

    Sean Cullen receives funding from the Engineering Design and Physical Sciences Research Council (EPSRC). This project specifically was funded through the Brunel Research Interdisciplinary Lab (BRIL). He is affiliated with the Space 4 All community.

    Ezgi Merdin Uygur receives funding from the Marketing Trust and the British Academy / Leverhulme.

    Vanja Garaj currently receives funding from Engineering and Physical Sciences Research Council (EPSRC), Science and Technology Facilities Council (STFC) and Research England.

    ref. Recruiting the world’s first disabled astronaut doesn’t mean space travel is inclusive – here’s how to change that – https://theconversation.com/recruiting-the-worlds-first-disabled-astronaut-doesnt-mean-space-travel-is-inclusive-heres-how-to-change-that-242397

    MIL OSI – Global Reports

  • MIL-OSI Global: Japan election: voters took aim at an untrustworthy government beset by scandal

    Source: The Conversation – UK – By Julie Gilson, Reader in Asian Studies, University of Birmingham

    Japan’s ruling Liberal Democratic party (LDP) suffered a severe blow on October 27 when, alongside its smaller coalition partner, Komeito, it lost its majority in a snap general election. The ruling coalition took 215 seats, fewer than the 233 required, with the centre-left opposition Constitutional Democratic party making big gains.

    Prime Minister Shigeru Ishiba called the election after winning his bid for party leadership in September. He had hoped to cement his position and draw a line under the tenure of his predecessor, Fumio Kishida, who had stepped down earlier that month amid a string of corruption scandals and public discontent over the rising cost of living.

    Ishiba has admitted that voters, who turned out in their third-lowest numbers in Japan’s post-war era, have dealt the LDP a “severe judgment”. But he has vowed to continue ruling the country.

    For its part, the opposition is not unified and therefore not in a position to offer a viable alternative. However, the ability of Ishiba’s government to push through the changes it needs to win back voter support will be severely restricted if the LDP fails to enter into coalition or garner key allies on particular issues.

    The LDP sits at the heart of the so-called “1955 system”, which has seen the party retain almost uninterrupted government control since the end of the second world war. But recent events have rocked Japanese politics.

    At the end of 2023, the public became aware of funding scandals involving dozens of LDP politicians. They were found to have diverted over ¥600 million (£3 million) of campaign donations into slush funds without recording the transactions as they were legally required to do.

    These scandals involved cabinet ministers and close allies of Kishida, who had already faced criticism over their links with the controversial Unification church. The church, whose members are commonly known as the Moonies, has been called a “dangerous cult” by its critics and is accused of exploiting its members financially.

    Japan’s former prime minister, Shinzo Abe, was shot dead in July 2022 by a man who said he held the church responsible for bankrupting his family. Abe was not a member of the church, but his grandfather was a key figure in its establishment in Japan in the 1950s. Kishida ordered party members to end their ties with the church in the aftermath of Abe’s assassination.

    These scandals have taken place against the backdrop of rising prices, stagnant wages and a generally sluggish economy. Consumer price inflation accelerated to 3% in August, a ten-month high. The dreary outlook contributed to voter disillusionment.

    According to a survey by Tokyo-based news agency Kyodo News, the approval rating of Ishiba’s cabinet fell to 32.1% after the vote, from its pre-election rating of 50.7%.

    The electorate has expressed its doubt that a new government could end the distrust caused by the scandals. Rebuilding this trust will only become harder as the yen continues to fall, and Japan’s economic uncertainty, ageing population, and disaffection among young voters persist.

    Regional insecurity

    The electoral body blow could also weaken Japanese foreign policy, with China emerging as the main beneficiary. To its democratic allies, a stable Japan is crucial for securing geopolitical stability in a region that also includes a dominant China, a belligerent Russia and a nuclear-armed North Korea.

    The LDP has traditionally always had a hawkish foreign policy stance. And in recent decades it has moved towards a desire to revise Japan’s “pacifist” constitution in favour of enabling the military to take a more flexible approach to security threats.

    Kishida was lauded abroad for his foreign policy, having proposed increases in the defence budget and more cooperation with the US in the Indo-Pacific region. And Ishiba has previously advocated for an “Asian Nato” to counter China. He has even visited Taiwan’s capital city, Taipei – much to Beijing’s disapproval.

    At the same time, Komeito’s more conservative position on foreign policy has supported an approach towards building diplomatic bridges with China. But should the LDP enter into coalition with the right-wing Japan Innovation party, which is a possibility given it won 38 seats in the recent election, a more assertive stance towards China may arise.

    Led by politician Nobuyuki Baba, the party supports the revision of Japan’s constitution and an increase in defence spending as a means of countering China’s regional influence.

    That said, a prolonged period of incapacitated politics within Japan presents a good opportunity for China to escalate its incursions into Japanese airspace and military manoeuvres around Taiwan. Japan’s leadership now needs to get its house in order quickly if the balance of security in the Indo-Pacific is to be maintained.

    Julie Gilson does not work for, consult, own shares in or receive funding from any company or organisation that would benefit from this article, and has disclosed no relevant affiliations beyond their academic appointment.

    ref. Japan election: voters took aim at an untrustworthy government beset by scandal – https://theconversation.com/japan-election-voters-took-aim-at-an-untrustworthy-government-beset-by-scandal-242406

    MIL OSI – Global Reports

  • MIL-OSI Global: How the state of our oceans is intrinsically linked to human health – new report

    Source: The Conversation – UK – By Edward H. Allison, Director of Science and Research, WorldFish, CGIAR System Organization

    eedafizie/Shutterstock, CC BY-NC-ND

    A new study published in the journal OneEarth explores how marine biodiversity conservation, human health and wellbeing are connected. The results suggest that marine protected areas can be good for both planet and people. These areas of the ocean are legally recognised by governments as being important for marine conservation. They are protected by putting limits on human activity within and around them.

    Once a government declares a marine protected area, you usually can’t live in it, fish, build a beach resort, start a fish farm or drill for oil in it. The rules vary from place to place, but the idea is to allow nature to flourish by limiting human activity as much as possible.

    With plans to expand ocean protection under the UN-endorsed biodiversity plan’s “30×30” target (which aims to protect 30% of the world’s land and oceans by 2030), it’s important to know how this will affect people as well as nature.

    The study was conducted by the conservation charity World Wide Fund for Nature, Harvard Institute of Public Health and Duke University’s marine laboratory. The team, led by marine conservation scientist Daniel Viana, reviewed all the scientific articles written since 1973 on marine protected areas and their impacts on people.

    They found that, for 234 marine protected areas across the world that have been closely monitored, more than 60% showed improvement in both nature conservation and human wellbeing.


    Swimming, sailing, even just building a sandcastle – the ocean benefits our physical and mental wellbeing. Curious about how a strong coastal connection helps drive marine conservation, scientists are diving in to investigate the power of blue health.

    This article is part of a series, Vitamin Sea, exploring how the ocean can be enhanced by our interaction with it.


    The study included marine protected areas that do allow “sustainable use” through managed and selective fishing activities. These are fishing methods, such as using a hook and line or a fish trap, that don’t cause physical damage to delicate habitats like coral reefs.

    The paper suggests that in most cases, investing in marine protected areas directly benefits the health and livelihoods of people who live near them. Increased harvests of fish and other aquatic foods, such as shellfish and seaweeds, are usually the source of the benefits. Fisherfolk’s incomes increase and community access to nutrient-rich aquatic food improves.

    Sustainably caught fish is a vital source of protein for so many people around the globe.
    M_Kaempfer/Shutterstock, CC BY-NC-ND



    Read more:
    Targets to save 30% of the ocean by 2030 aren’t being met, new report reveals


    The benefits of marine protection for fishing-based livelihoods are largest in small island states that have big marine protected areas, such as Bonnaire, Palau and the Cook Islands, where more than 95% of fish catches are associated with area-based conservation measures.

    Despite ample evidence that marine protection improved access to aquatic food, the authors found surprisingly few studies that directly measured the impact to human nutrition. Only three out of the 237 studies reviewed had studied how creating marine protected areas affected the diets of people living around them. Only one study, in the Philippines, made the link between diets and health outcomes, because, when access to fish in diets improved due to marine conservation, there were fewer stunted children from surrounding communities.

    Plenty more nutrients in the sea?

    Our continents and islands are surrounded by seas, lakes, rivers and floodplains that are populated by edible plants and animals rich in vitamins, minerals and fatty acids. These micronutrients from aquatic foods are highly bioavailable (easily absorbed by the body). If sustainably harvested and made available to nutritionally vulnerable people, they could prevent malnutrition among millions of coastal people.

    The new report has quantified the micronutrient contributions to human diets from the aquatic foods that flourish when marine protected areas are set up. It combines data on the nutrient composition of all the aquatic foods harvested in and around marine protected areas, with fish catch data from the surrounding areas.

    The existing marine protected area network supports 14% of the global supply of six key micronutrients from marine fishing. This is achieved by protecting only 8% of the world’s oceans. By allowing marine life to grow abundantly inside protected areas, nearby fish populations are replenished. So, by conserving marine wildlife, protected areas help to sustain fish and shellfish stocks.

    That means bigger catches, more income from fishing or tourism, and more food. More nutrients means better health. This applies both to marine protected areas with a strict no-take zone, where any form of fishing is banned, and those that allow regulated fishing.

    As populations increase, demand for aquatic food rises. Wild harvests are being supplemented by aquaculture and mariculture – these are freshwater and marine equivalents to growing crops and livestock on land. Over half of the aquatic foods consumed directly by humans are now produced from aquaculture, much of it in inland waters rather than the sea.

    But in many countries, particularly island and coastal nations in the developing world, harvesting wild food from marine ecosystems remains crucial to nourishing the over 3 billion people who get more than 15% of their animal source proteins from aquatic foods.

    Seafood is a rich source of vitamins, minerals and fatty acids.
    WhiteYura/Shutterstock, CC BY-NC-ND

    Despite their potential to address global micronutrient nutrition, aquatic foods have, until recently, been underrepresented in policies and programmes to end hunger and malnutrition. But with data on the nutritional composition of the world’s fish species now available, studies like this can advance an approach called “nutrition-sensitive fisheries and aquaculture”: Instead of fishing to maximise catch or profit, fisheries could be managed to optimise their contribution to human nutrition.

    Linking ocean conservation with human health is an exciting idea but there are gaps in the research. It’s not clear who benefits when income from tourism and fishing increases, or whether increased catches get to those that need it most. In the Maldives for example, more than 80% of reef fish are consumed by tourists, not locals.

    Trying to solve malnutrition with marine protected areas is going to be challenging. Many marine protected areas are not effectively managed. By contrast, 77% of catches from the world’s fisheries come from stocks that are managed sustainably, though they have little room for expansion to meet rising demand. Aquaculture can do that, but the sector is still moving towards sustainability.

    Many key threats to marine ecosystems and wild fisheries, such as climate change and pollution, are not effectively dealt with by local marine habitat protection alone. Despite these challenges, this study highlights that nature-human relationships can be regenerative, rather than exploitative.



    Don’t have time to read about climate change as much as you’d like?

    Get a weekly roundup in your inbox instead. Every Wednesday, The Conversation’s environment editor writes Imagine, a short email that goes a little deeper into just one climate issue. Join the 40,000+ readers who’ve subscribed so far.


    Edward H. Allison currently receives funding from Canada’s International Development Research Center AQUADAPT programme for work on climate adaptive nature-based aquaculture in South East Asia, from the UN Food and Agriculture Organization for work on Implementing ecosystem-based management in S and SE Asia arnd from the multi-donor Trust Fund to the Consultative Group on International Agricultural Research for work on aquatic food systems.

    ref. How the state of our oceans is intrinsically linked to human health – new report – https://theconversation.com/how-the-state-of-our-oceans-is-intrinsically-linked-to-human-health-new-report-242245

    MIL OSI – Global Reports

  • MIL-OSI Global: Did a Canadian developer really invent bitcoin? A new HBO show explores an intriguing theory

    Source: The Conversation – Canada – By Jeremy Clark, Associate Professor, Information Systems Engineering, Concordia University

    The true identity of the founder of bitcoin has always been a mystery. (Shutterstock)

    In 2008, someone using the pseudonym Satoshi Nakamoto published the design of the cryptocurrency bitcoin, proposed the initial code and was active online for just under two years. In this time, they helped develop the code, answer questions and promote the project. Then, claiming to busy with new things, Nakamoto left working on bitcoin and was probably never heard from again.

    HBO’s 2024 documentary Money Electric: The Bitcoin Mystery finds director Cullen Hoback looking for the real Nakamoto, motivated by bitcoin being “embraced by nation states” and “incorporated into 401(k)s.”




    Read more:
    Bitcoin turns ten – here’s how it all started and what the future might hold


    The real Nakamoto?

    Several attempts to unmask Nakamoto have been made before. Previous theories suggest that the elusive developer is Irish graduate student Michael Clear, Japanese-American systems engineer Dorian Nakamoto or one of several cypherpunks who worked on predecessors to bitcoin: Hal Finney, Nick Szabo or Adam Back.

    Hoback confronts the man he suspects of being Nakamoto on camera in the film’s climax: Peter Todd, a software developer from Toronto. On film, Todd alternates between joking about being Nakamoto and calling the theory ludicrous, perhaps necessitating him to make an unequivocal denial in the press after it aired.

    The trailer for HBO’s ‘Money Electric.’

    The documentary is entertaining, but does it play it fast and loose? I would draw attention to three things that deserve further thought.

    Online breadcrumb trail

    While stopping short of claiming to have conclusively identified bitcoin’s creator, Hoback suggests something Todd once said to Nakamoto online was a slip up.

    The background is this: with bitcoin, users leave tips to have their transactions processed. If the tip is too low, the computers running bitcoin will refuse to process it and the transaction will sit in bitcoin purgatory. Worse, bitcoin users who make this mistake cannot increase the fee without it looking like an attack on the system.

    In an online post, Nakamoto posts that transactions could be declared safe if they only changed the amount of the fee.

    Not long after, Todd chimes in that this is impossible with how bitcoin transactions work. The increased fee has to come from somewhere, namely a decrease in the amount paid out, which changes the transaction. Todd’s message is short: “Of course, to be specific, the inputs and outputs can’t match *exactly* if the second transaction has a transaction fee.”

    Hoback ponders if maybe Nakamoto meant to correct himself, but somehow accidentally used his real account.

    As the documentary recounts, Todd is smart, has developer experience and had been discussing digital cash online since he was a teenager. Todd would eventually be the one to implement the feature Nakamoto described, albeit with a fix to the issue he pointed out.

    The theory plays out well on film but leaves out a few considerations.

    Early bitcoin enthusiasts were a self-selecting group, and most were as technically minded as Nakamoto or Todd. This technical background is niche but not rare: more than 100,000 computer science students graduate annually in the United States, while there are over 500,000 certified security experts. And there are many equally capable people who are neither of these things.

    Given Hoback’s evidence for Todd is circumstantial, the weight shifts to Todd’s reaction on camera when Hoback outlines his theory: a mix of bemusement, mockery and indignation. The film frames the reaction as incriminating, while others caution against reading anything into it.

    Enter Ethereum

    Bitcoin is maintained by an open group of volunteered computers (whose operators are paid in new bitcoin for the work of validating transactions and storing them on a ledger called the blockchain) where no one is in charge, and yet maintains high security.

    Early bitcoin enthusiasts saw the potential for bitcoin’s blockchain technology to handle more than financial transactions, but the developers helming bitcoin (including Todd) thought it would be best if bitcoin stayed in its lane.

    Some bitcoin enthusiasts in Toronto then banded together and launched Ethereum. Led by 21-year-old Vitalik Buterin, Ethereum provides a platform where anyone can run their code on a blockchain simply by paying a fee and pushing a button. The code could be anything from a new digital currency to sophisticated financial technology.

    In Hoback’s documentary, many of the interviewees view bitcoin and its developers as competitors and antagonists of Ethereum.

    Ethereum gets only about two minutes of screentime, dominated by Buterin rapping about Ethereum on the mainstage of a conference and being ribbed for his hat’s safari flaps.

    Hoback’s documentary emphasizes Ethereum’s scam tokens but overlooks the innovative financial services that captured US$64 billion of assets in 2021, as well as its advancements in areas like efficiency and cryptography.

    Ironically, it is Ethereum technology that runs crypto-betting platform Polymarket, which hosted a US$44 million betting pool on who would be named as Nakamoto in Hoback’s film before it aired.

    “Polymarket turned Money Electric into a sporting event,” Hoback enthused. “Even I’m refreshing the betting pool to see how high the total volume gets.”

    The end of privacy?

    In his 2014 documentary, Terms and Conditions May Apply, Hoback did show he is willing to tackle social concerns that might seem a little dry or academic, such as privacy rights in a digital age.

    He picks up this thread again in Money Electric, embedding an earnest message about the potential privacy and surveillance implications of governments — including Canada, the United States and 130 other countries — launching central bank digital currencies (CBDCs), something my research also draws attention to.

    In theory, the technology underlying bitcoin can be expanded to provide a CBDC system as private as paper cash. However it will take a strong political will to get there.

    Jeremy Clark receives funding from the National Sciences and Engineering Research Council (NSERC), Raymond Chabot Grant Thornton and Autorité des Marchés Financiers.

    ref. Did a Canadian developer really invent bitcoin? A new HBO show explores an intriguing theory – https://theconversation.com/did-a-canadian-developer-really-invent-bitcoin-a-new-hbo-show-explores-an-intriguing-theory-241750

    MIL OSI – Global Reports

  • MIL-OSI USA: Hickenlooper, Bennet, Neguse, Pettersen, Polis Announce $129 Million for Colorado Rail Projects 

    US Senate News:

    Source: United States Senator John Hickenlooper – Colorado
    Four Colorado projects awarded funding under the Consolidated Rail Infrastructure & Safety Improvements (CRISI) Grant Program
    WASHINGTON – Today, U.S. Senators John Hickenlooper and Michael Bennet, U.S. Representatives Joe Neguse and Brittany Pettersen and Governor Jared Polis announced four Colorado rail projects will receive a total of $129.5 million in federal funds. The Colorado Department of Transportation (CDOT), Colorado State University Pueblo, San Luis Central Railroad Co., and OmniTRAX will all receive funding as part of the Consolidated Rail Infrastructure & Safety Improvements (CRISI) Grant Program. Earlier this year Hickenlooper, Bennet, Neguse and Pettersen urged the U.S. Department of Transportation to fund CDOT’s project along the Front Range. Hickenlooper also urged the department to fund the CSU Pueblo and OmniTRAX projects.
    “From freight in the San Luis Valley to passengers on the Front Range and beyond with CSU Pueblo’s research, rail isn’t just a part of our past, it’s a big part of our future, too,” said Hickenlooper. “That’s the case we made to Secretary Buttigieg for this funding and this is just the start.”
    “Colorado’s railways are vital to connect our communities and get resources to markets across the country. That’s why I ensured the U.S. Department of Transportation understood how critical this funding is for our state’s transportation infrastructure,” said Senator Michael Bennet. “I’m glad to have helped secure these investments in our railways’ safety, efficiency, and reliability across the state. ”
    “After years of working to secure federal support for the Front Range Passenger Rail Project, I am excited to see the Department of Transportation heed our calls and commit to modernizing Colorado’s passenger rail system—not just for communities along the Front Range but for residents throughout the entire state. This is a once-in-a-generation investment in our passenger rail infrastructure, creating countless new opportunities for communities to connect, grow, and thrive—and we will continue to work together to ensure this momentum leads to lasting benefits for all Coloradans,” said U.S. House Assistant Minority Leader Joe Neguse.
    “Today, I am incredibly grateful to see this federal funding coming to Colorado to strengthen our railway systems, enhance safety, and modernize our infrastructure,” said Representative Brittany Pettersen. “After a train derailment in Boulder injured workers and put our communities at risk, I supported funding to reinforce public safety and restore trust in Colorado’s rail infrastructure. I’m pleased to see these federal dollars coming to our state to help ensure we have safe, reliable infrastructure for generations to come.”
    “Today’s grant will make freight rail traffic in some of our busiest growing communities safer quickly while providing critical building blocks for Passenger Rail.  This major funding will help achieve important priorities like complying with longstanding federal standards and improving the safety of rail crossings, which can be the sites of dangerous incidents. With more than $66 million in federal support from the Biden-Harris administration, the future of Colorado’s rail network is a clear priority for the federal government, as it should be. We thank Senators Hickenlooper and Bennet, Congressman Neguse and Congresswoman Pettersen, and our communities for their support of this important project,” said Governor Jared Polis.
    “Thanks to a unified effort with Governor Polis’ leadership, Colorado can speed ahead with important safety and operational upgrades that will make passenger rail possible along the Front Range. Our partners in the Congressional delegation and in communities across the state have been constantly supportive of this work, and I want to especially thank the technical team at CDOT that has made so much progress behind the scenes to get Colorado ready for this opportunity. The Biden Administration has recognized Colorado’s seriousness and the quality of our work to develop passenger rail, and I want to add my appreciation to their support with this grant and the resources it brings to our work,” said CDOT Executive Director Shoshana Lew.
    CRISI invests in railroad infrastructure projects that improve safety, support economic vitality, including through small businesses, create good-paying jobs with the free and fair choice to join a union, increase capacity and supply chain resilience, apply innovative technology, and explicitly address climate change, gender equity, and racial equity. For more information on CRISI, click HERE.
    Full details on the projects receiving funding are below:
    Recipient
    Project Title
    Project Description
    Amount Awarded
    Colorado Department of Transportation
    Modernizing Rail on the Front Range: PTC Installation, Siding, & Grade Crossing Safety and Operational Improvements
    This project will design, install, and test positive train control with a complementary siding on a portion of the Front Range Subdivision, along with several railroad crossings that could benefit from operational and safety improvements.
    $66,400,000
    OmniTRAX Holdings Combined, Inc.
    Transportation Investments for Employment and Safety, Phase 2
    The proposed project involves final design and construction activities to replace railroad ties on four OmniTRAX-owned short lines across four states – Alabama, Colorado, Georgia, and Washington.
    $50,570,400
    Colorado State University Pueblo
    Safety Assessment, Testing and Workforce Development for Hydrogen/Natural Gas Motive Power
    The proposed project involves research and development for studying green hydrogen and renewable natural gas-powered rail vehicles. The project aims to conduct safety experiments on the use of CH2/CNG-powered rail cars at the TTC facility.
    $11,671,781
    The San Luis Central Railroad Co.
    The San Luis Central Railroad Reconstruction Project: Ansel North
    The SLC corridor was built in 1913 with untreated wooden ties. The project will replace 6,000 deteriorated cross and 126 switch ties between mile posts 10.1 and 15.2.
    $1,077,000
    “Southern Colorado often represents a hard-working spirit leveraging the opportunity of innovation. This Department of Transportation CRISI grant emboldens that spirit, enabling CSU Pueblo, in partnership with the Southern Colorado Transportation Technology Center (SCITT), to contribute to the future of rail transportation through critical safety research in hydrogen and natural gas technologies. I am particularly proud of how this project will partner with our Engineering program at CSU Pueblo, utilizing the expertise here to create new pathways for our students and local workforce. This grant is more than research – it’s a valuable investment into Southern Colorado,” said CSU Pueblo President Armando Valdez.
    “TIES2 will be transformative for the communities served by Great Western Railway of Colorado and the regions served by OmniTRAX railroads in Georgia, Alabama, and Washington state,” said David Arganbright, OmniTRAX Senior Vice President. “OmniTRAX is proud to call Colorado home, and we are tremendously appreciative of all the work that Sen. Hickenlooper has done in Congress to champion Colorado’s railways and deliver the critical infrastructure investments that make strengthen our nation’s supply chains.”
    “The team at CXSL is very excited for this great news and look forward to getting to work on the improvements as soon as possible. The grant will assist in providing the much needed improvements to improve rail service to our customers and greatly reduce our risk for incidents due to track conditions,” said Timothy Bivens, General Manager of Colorado Pacific San Luis Railroad.
     

    MIL OSI USA News

  • MIL-OSI Global: Labour’s first budget: redistribution away from the rich after over a decade of Conservative rule

    Source: The Conversation – UK – By Martin Smith, Anniversary Professor of Politics, University of York

    Treasury/Flickr, CC BY-NC-ND

    Labour’s first budget in 14 years appears to have avoided the Halloween nightmares that many predicted. Yes, the overall tax burden is at its highest level since 1948, but this budget’s tax and spend distribution is such that it spreads both the pain and the benefit. Most of the pain has been focused on the well off, sparing others from increased taxation.

    This was, as is so often the case, a budget that was more about politics than economics.

    The political emphasis is naturally very different from the previous Conservative administration. Like the 1990s governments of Tony Blair, Labour is now focusing on improving the public sector rather than cutting taxes.

    The government claims there will be no return to austerity. Instead, Reeves’s budget is based not only on investment and growth, but education – which gets a 19% real-terms funding uplift. An extra £22 billion is also available for the NHS. Perhaps more importantly, there is an attempt to shift away from hospital-based provision to preventative approaches and community care.

    This is a budget centred on redistribution. Taxes are increasing for employers (through increased national insurance) and on inheritance tax and capital gains. Second home owners, non-doms and parents sending their children to private school will also be taxed more. Beneficiaries include those on a minimum wage, pensions and carers.


    Want more politics coverage from academic experts? Every week, we bring you informed analysis of developments in government and fact check the claims being made.

    Sign up for our weekly politics newsletter, delivered every Friday.


    It is, then, very much a Labour budget. It focuses on taking money from what may be called the upper middle class and the very well off, and spending it on the public sector, the lower paid and the worse off in society. Even the hinted-at increases on taxes around pensions and pension pots were not fulfilled for fear of alienating Labour supporting, public sector workers.

    Delivering a vision

    From this perspective, the budget can be regarded as a political success. It has done what Keir Starmer’s government has hitherto failed to do: set out a new, distinct agenda. Revising the existing fiscal rules to include both new stability and investment rules highlights Labour’s commitment to longer-term growth, which it hopes will secure the investment for renewed public services.

    This strategy harks back to the traditional social democracy that runs from Tony Crosland – one of the foremost Labour thinkers on a reformed social democracy – and the Labour governments of the 1960s to New Labour. Economic growth, rather than the radical redistribution of wealth associated with Labour leaders such as Michael Foot and Jeremy Corbyn, is Labour’s mechanism for enhancing public services and improving the position of the worst off in society.

    Budget 2024: a political document first, an economic plan second.
    Treasury/Flickr, CC BY-NC-ND

    But therein lies the rub. The initial market reaction was good. Much of Reeves and Starmer’s pre-budget spin was about making sure the markets remained calm. There was no repeat of the Liz Truss and Kwasi Kwarteng mini-budget debacle, where unfunded tax cuts led to the unstable economic conditions and ultimately Truss’s downfall.

    The whole of the Labour government strategy is based on modest but consistent economic growth between a high of 2% and a low of 1.5% between now and 2030. But, of course, economic growth is very difficult to predict and dependent on conditions that the government does not control.

    Just this week, Israel’s decision not to target Iranian oilfields led to a 20% drop in oil prices. But any intensification of war in the Middle East could see that situation rapidly reversed.

    So while Labour’s promise to increase capital spending and greater investment in science, research and development is important for growing the economy, it is only one factor and others may thwart Labour’s growth plan.

    There is though one important lesson from history. All – and that really is all – Labour governments have ended their time in office amidst a financial crisis. Often, it is not directly of their own making, but the plight of the economy has subsequently undermined their original spending plans.

    Reeves’s first budget has seen clear benefits for particular sectors. The adaptation of the fiscal rules creates new opportunities for growth through borrowing, capital spending and investment. Yet whether it can trigger the scale of economic takeoff needed to overcome the backlog of investment in public services is to be seen. This may well prove to be the key factor in determining both the longevity and legacy of this Labour government.

    Martin Smith receives funding from Leverhulme Major Research Fellowship and Nuffield Foundation.

    Dave Richards receives funding from two projects related to this article: 1. ‘Public Expenditure, Planning and Control in Complex Times’ – Nuffield Foundation, https://sites.manchester.ac.uk/public-expenditure-planning-and-control/ 2. ‘The UK Productivity-Governance Puzzle: Are UK’s Governing Institutions Fit for Purpose in the 21st Century?’ – The ESRC Productivity Institute https://www.productivity.ac.uk/people/professor-david-richards/

    Sam Warner receives funding from the Nuffield Foundation.

    ref. Labour’s first budget: redistribution away from the rich after over a decade of Conservative rule – https://theconversation.com/labours-first-budget-redistribution-away-from-the-rich-after-over-a-decade-of-conservative-rule-242548

    MIL OSI – Global Reports

  • MIL-OSI United Kingdom: expert reaction to study on sugar rationing in utero and early life reducing the risk of chronic disease in adulthood using post-WWII data

    Source: United Kingdom – Executive Government & Departments

    A study published in Science looks at sugar rationing in the first 1000 days of life and the risk of chronic disease in adulthood. 

    Dr Hilda Mulrooney, Reader in Nutrition and Health, London Metropolitan University, said:

    “This is a really interesting and timely paper, given the currently high intakes of sugar in the UK population, and prevalence of chronic disease including Type 2 diabetes and hypertension. 

    “The potential for diet in utero to impact on long-term health risks has long been recognised, and there are a number of plausible mechanisms to explain how these may occur. In this study, the authors used data from what could be considered a natural experiment – rationing in response to World War 2. By comparing individuals exposed and not exposed to sugar rationing in utero and in early childhood, a significant effect was seen. Those exposed to rationing had significantly lower risks of Type 2 diabetes (35% lower) and hypertension (20% lower), compared to those who did not. Early childhood was especially important; only a third of the increase in risk for both type 2 diabetes and hypertension was explained by in utero exposure. This highlights the potential for early childhood diet as a risk factor for chronic disease. Given the high levels of sugar in foods and drinks aimed at toddlers and young children, this is of concern.

    “The study cannot demonstrate causality; it is not possible from this sort of study design. Nonetheless it is a strong study, with several potential confounding factors taken into account and large numbers of participants (38,155 exposed to rationing and 22,028 not exposed to rationing). The groups were similar in relation to gender, race, family history of diabetes and cardiovascular disease (for which hypertension is a strong risk factor), and genetic scores calculated for obesity (which could confound for both type 2 diabetes and hypertension). In addition to sugar intakes immediately and markedly rising after rationing of sugar ended, risk of obesity was also significantly higher in those not exposed to rationing in utero and the first year of life, compared with those who were. 

    “This study inevitably has weaknesses, due to its type and reliance on historical data. Changes to sugar intakes were unlikely to be the only changes that occurred to participants at that time. The authors  took as many factors as they could into account and the message is clear – exposure to high intakes of added sugar in utero and early childhood is a significant risk factor for chronic disease. This suggests that action to reduce the sugar content of foods and drinks aimed at or attractive to children is needed. This  will not entirely reduce the risk, since diet in pregnancy is what enables in utero exposure – so action on all foods and drinks high in added sugar is needed. However this will have to be approached with care – simply replacing sugar with sweeteners will not enable the population to reduce their preference for sweet tastes.”

     

    Jerusa Brignardello, Lecturer in Dietetics & Nutrition, Oxford Brookes University, said:

    Does the press release accurately reflect the science?

    “Yes, the press release is aligned with the conclusions and findings to the study. This press release emphasises in the importance of the sugar restriction during the first 1000 days of life as an early dietary intervention for the reduction of hypertension and diabetes risk.”

    Is this good quality research?  Are the conclusions backed up by solid data?

    “This is an interesting retrospective piece of research that explores the consequences of early of sugar restriction during early life and comparing the effects of sugar consumption in the same age group after the rationing of food was lifted in the post-war UK. However, results should be interpreted with caution, as nutritional environments from the 1950s differ significantly from those of today.

    “The information was obtained from the UK- Biobank with 60183 participants. However, the Biobank cohort is not nationally representative of the population and tend to represent a part of the population that was healthier and wealthy. Nonetheless, the quasi-experimental design of the exposure conditions makes this study very rigorous to study the sugar exposure in humans.”

    How does this work fit with the existing evidence?

    “Current evidence suggests that early exposure to sugar during pregnancy and early life may impact neonatal metabolism, obesity risk, and taste perception, which may later influence food choices and the risk of other chronic diseases. This study supports the findings related to chronic diseases and contributes to the “fetal origins hypothesis” described by British physician and epidemiologist David Baker in the 90s.”

    Have the authors accounted for confounders?  Are there important limitations to be aware of?

    “Yes, the authors have worked on the limitations inherent in studying a cohort like this. It is important to be aware that food environments and dietary patterns in the 50s were very different compared to the current food environment. In addition, lifestyle during those years was not the same as today, and obesity was not considered for statistical purposes as a potential variable to study public health.

    “Therefore, the risk found in the UK-Biobank cohort that was exposed to sugar rationing may be different if that is compared to other populations given the differences in lifestyle, dietary habits, food environment, and obesity prevalence. Consequently, the extrapolation of the results presented in the work of Gracner et al. should be interpreted with caution, for example, if these results will be used to build machine learning models for risk predictions for the current population. However, these results contribute to the “Baker hypothesis” or “Fetal origins hypothesis” showing how a simple nutritional intervention as cutting sugar during this crucial period of pregnancy and early life affect in the reduction of risk of diabetes and hypertension in later life.”

    What are the implications in the real world?  Is there any overspeculation?

    “As mentioned previously, the food environment, lifestyle, and physical activity are very different from those in their 50s. Therefore, the results found in this research should be a call for attention for women in the stages of preconception, pregnancy, and parents of children in early life. In addition, this should reinforce the actions of policymakers for the promotion of low sugar intake during these critical life stages in parents and children. Finally, the food industry should consider reformulating products targeted at these groups in light of the evidence, prioritising the well-being of future generations.

    “I do not believe there is overspeculation in this article, as it has undergone peer review, meaning that multiple academics have evaluated the research, including its methodology, results, discussion, and conclusions.”

     

    Dr Katie Dalrymple Lecturer in Nutritional Sciences, Kings College London said:

    “This study provides further epidemiological evidence which supports the Developmental Origins of Health and Disease (DOHaD) hypothesis. DoHaD suggests that certain environmental influences during critical periods of growth and development during early life may have significant consequences on a child’s long-term health. Given the complexity of this research question, the study relies on observational data and an event study design to draw meaningful conclusions of the relationship between nutrition in early life and the development of chronic diseases. Whilst it is important to consider confounding factors which may have occurred between the exposure and the outcome and potential bias of the Biobank cohort, the results are consistent with existing DoHaD literature, and they support the notion of public health initiatives which focus on sugar reduction.”   

    Amanda Adler, Professor of Diabetic Medicine and Health Policy from the University of Oxford’s Radcliffe Department of Medicine said:

    “The investigators take advantage of the ‘natural experiment’ of post-war food rationing to test the theory that exposure to sugar rationing in utero and in early childhood prevents or delays the onset of type 2 diabetes and hypertension years later.

    “The investigators observed that people conceived during rationing indeed had lower rates of disease when compared to people conceived after rationing ended.

    “But, we still don’t really know if the children less likely to get diabetes later in life were indeed the ones not exposed to sugar in utero or after birth – even in a setting of rationing.

    “It may be that at the same time rationing ended and people consumed more sugar, they also changed other habits becoming, for example, less physically active.  So, this may have influenced in part their risk for diabetes later in life. 

    “It’s intriguing and entirely possible that a lower exposure to sugar in utero via the mother would lead to life-long benefits. 

    “This study is an open invitation to clinical trialist to clarify the ‘right’ levels of sugar to add to the diet for pregnant or lactating women, and for their infants.”

    Exposure to sugar rationing in the first 1000 days of life protected against chronic disease’ by Gracner et al. was published by AAAS in the journal Science at 18:00 UK time on Thursday 31st October.

    DOI: 10.1126/science.adn5421

    Declared interests

    Dr Hilda Mulrooney “In terms of conflicts, I am a committee member of the Obesity Group of the British Dietetic Association, a committee member of the European Specialist Dietitians Network for Obesity and a Council member for Public Health to the Nutrition Society. I am not paid by any of these organisations and not representing them in these comments.”

    Jerusa Brignardello “In 2013 I was awarded scholarships from Kraft Foods to attend to the Young Global Nutrition Leader in the International Unions of Nutrition Societies and International Nutrition Foundation. I worked as International Nutrition Consultant for the World Food Programme at United Nations in the Latin American and Caribbean Bureau between 2013 and 2014 . I have worked in Nutrigenomix which is a company for nutritional genetic testing based in Canada between 2012 and 2017. Also, as a clinical trial coordinator for Nestle Switzerland in 2010 and as consultant for Nestle Chile doing activities related to science communication in gut health topics in 2024. In 2018 I received a funding from the American Egg Board from USA to do research in food biomarkers, while I studied at Imperia College London- UK. I am not aware about significant industry funding in my department at Oxford Brookes University. I do not have any conflicts of interest related to this research for my own research.”

    Dr Katie Dalrymple “I worked for Danone for 4 years (2012-2016) before I did my PhD.”

    Amanda Adler “No conflicts of interest to declare.”

    MIL OSI United Kingdom

  • MIL-OSI Canada: Funding to support 14 First Nations, Inuit and Métis documentary heritage projects

    Source: Government of Canada News

    Library and Archives Canada is pleased to announce the recipients of the 2024–2025 funding cycle of its Listen, Hear Our Voices initiative.

    Gatineau, Quebec, October 31, 2024

    Library and Archives Canada (LAC) is pleased to announce the recipients of the 2024–2025 funding cycle of its Listen, Hear Our Voices initiative. In total, 14 First Nations, Inuit and Métis organizations will share $912,963 to support their efforts to digitize and preserve language and culture materials and to increase their digitization and preservation capacity.

    The recipients were selected based on the recommendations of a review committee, external to LAC and consisting of First Nations, Inuit and Métis heritage professionals from across Canada.

    Indigenous organizations receiving funding through the Listen, Hear Our Voices initiative (2024–2025):

    • Cold Lake First Nations
    • Native Counselling Services of Alberta
    • Ktunaxa Nation Council Society
    • Cheslatta Carrier Nation
    • Gwich’in Tribal Council
    • Inuvialuit Cultural Centre Pitquhiit-Pitqusiit
    • Inuit Broadcasting Corporation
    • Woodland Cultural Centre
    • Munsee Delaware Language and History Group, c/o Munsee-Delaware Nation
    • Chippewas of Rama First Nation
    • La Boîte Rouge VIF
    • First Nations University of Canada
    • Kinistino Metis Local 43
    • Council of Yukon First Nations—Yukon Native Language Centre

    Project titles and funding amounts for 2024–2025 contribution funding recipients can be found on the Listen, Hear Our Voices web page.

    Heritage organizations play a vital role in preserving Canada’s local memory. LAC is proud to contribute to documenting, preserving and making accessible to the public a memory that reflects the diversity of experiences and cultures in Canadian society.

    About Library and Archives Canada

    The mandate of Library and Archives Canada is to preserve the documentary heritage of Canada for the benefit of present and future generations and to be a source of enduring knowledge accessible to all, thereby contributing to the cultural, social and economic advancement of Canada. Library and Archives Canada also facilitates cooperation among communities involved in the acquisition, preservation and dissemination of knowledge and serves as the continuing memory of the Government of Canada and its institutions.

    MIL OSI Canada News

  • MIL-OSI Global: Sudan’s civil war has left at least 62,000 dead by our estimate − but the true figure could be far higher

    Source: The Conversation – USA – By Sarah Elizabeth Scales, Post-Doctoral Researcher, Department of Environmental, Occupational, and Agricultural Health, University of Nebraska Medical Center

    The ongoing war in Sudan has often been overlooked amid higher-profile conflicts raging across multiple continents. Yet the lack of media and geopolitical attention to this 18-month-long conflict has not made its devastation in terms of human lives any less stark.

    Since fighting broke out in April 2023 between the Sudanese Armed Forces and the paramilitary Rapid Support Forces, both of which had been part of a power-sharing military government, the country has seen the displacement of more than 14 million people and the carving up of the country by geography and ideology.

    And while we may never know the exact death toll, the conflict in Sudan is certainly among the deadliest in the world today.

    As scholars of public health, conflict and human rights and Sudanese-American health workers, we are keenly aware of how fraught it can be to estimate mortality in war for a slew of practical and political reasons. But such estimates are of critical importance: They allow us to understand and compare conflicts, target humanitarian aid for those still at risk, trigger investigations of war crimes, bear witness to conflict and compel states and armed groups to intervene or change.

    The difficult work of counting the dead

    A profound humanitarian crisis is occurring in Sudan, characterized by ethnic cleansing, mass displacement, food scarcity and the spread of disease, complicated further by flooding in the northern states.

    Considering a death toll in such a conflict includes counting not only those who are killed as a direct result of violence – itself a difficult thing to determine in real time – but also those who have died by conflict-exacerbated factors, such as the absence of emergency care, the breakdown of vaccination programs and a lack of essential food and medicine. Estimating this latter death toll, called indirect mortality, presents its own challenge, as the definition itself varies among researchers.

    In congressional testimony, U.S. special envoy to Sudan Tom Perriello recognized the estimation challenges when noting there had been anywhere between 15,000 and 150,000 deaths in Sudan – an exceedingly wide range that was attributable, in part, to the complexity of determining indirect mortality.

    Armed Conflict Location and Event Data (ACLED), a nonprofit specializing in conflict-related data collection, has recorded an average of more than 1,200 direct conflict deaths per month in Sudan, with nearly 19,000 deaths in the first 15 months of the conflict. This figure is similar to the 20,000 deaths estimated by the Sudan Doctors Union and the 19,000 figure used by the Sudan Protection Cluster, a centralized group of U.N. agencies and NGOs that used World Health Organization data.

    ACLED sources its estimates of deaths from traditional media, reports from international NGOs and local observers, supplemented by new media such as verified Telegram and WhatsApp accounts. The Sudan Doctors Union, on the other hand, gives on-the-ground estimates of conflict deaths.

    When available, distinct data sources such as surveys, civil registers and official body counts can make an estimation more accurate. However, this data is often available only in retrospect, after the cessation of conflict. It is therefore critical to use both the available data and precedents from previous conflicts to capture a reasonable estimate of the human costs of an ongoing conflict.

    Internally displaced Sudanese children in Port Sudan, Sudan, on Jan. 3, 2024.
    Omer Erdem/Anadolu via Getty Images

    A 2010 article in The Lancet estimated that there are 2.3 indirect deaths for every direct conflict death, based on data from 24 small-scale surveys conducted in Darfur from 2003 to 2005. As such, using ACLED’s data of 18,916 direct deaths, we estimate that in the current Sudan conflict, there are an additional 43,507 indirect deaths – or more than 62,000 total deaths.

    We believe our estimate is very conservative. When estimating mortality in the ongoing conflict in Gaza, a different group of scholars, also writing in The Lancet, used a multiplier of four indirect deaths for every direct death to estimate the overall mortality there.

    Meanwhile, a report from the Geneva Declaration Secretariat showed an average of 5.8 indirect deaths for every direct death across 13 armed conflicts from 1974 to 2007.

    Using that latter multiplier, the number of indirect deaths in Sudan would jump to nearly 110,000 – meaning the total deaths in the region amount to 130,000 – double our estimate.

    This range is wide, but it acknowledges how difficult it can be to estimate indirect deaths and how they can vary significantly with the shape of a conflict.

    The Sudanese conflict in context

    For all the tremendous loss of life these numbers reflect, they surely underestimate the true human costs of the conflict.

    Sudan already had a fragile and underfunded health system before the fighting started. And compared with other ongoing conflicts such as in Gaza and Ukraine, there was already a more precarious baseline, with higher child mortality and lower life expectancy.

    Since the war in Sudan began, there have been consistent reports of mass killings, forced disappearances, sexual violence, deliberate blocking of food and medicine, and other forms of violence against civilians.

    Much of the violence is ethnically targeted, and the Darfur region – where a full-scale famine has been declared – has suffered disproportionately.

    The destruction of civilian infrastructure and interrupted aid mechanisms are preventing medicine, food, clean water and vaccinations from getting to in-need populations.

    Health care workers and facilities, not only in at-risk Darfur but also throughout the country, have been the target of attacks. Nearly 80% of medical facilities have been rendered inoperable. And at least 58 physicians have been killed, in addition to the many that were targeted in previous crises.

    Given the persistent targeting of health care systems and restricted access to humanitarian corridors, indirect deaths in Sudan are likely to grow as hospitals shut down, even in the capital Khartoum, due to bombardments, ground attacks and a lack of critical supplies.

    The costs for Sudanese children are especially alarming. Thirteen children die per day in Zamzam camp in North Darfur, according to Doctors Without Borders, mostly due to undernutrition and food scarcity.

    And nearly 800,000 Sudanese children will face severe, acute malnutrition through 2024, a condition that requires intensive care and supplemental nutrition merely to prevent death. Even before the conflict, children were severely threatened by a lack of access to care, including basic preventive care such as early immunization.

    Finally, the transmission of communicable diseases thrives in conflicts like the one in Sudan, where there has been widespread population displacement, malnutrition, limited water and sanitation, and lack of appropriate sheltering. In August, a cholera outbreak led to a spiking death rate of more than 31 deaths per 1,000 cholera cases. And instances of such disease effects are likely underestimates in a country lacking health care penetration and monitoring.

    The limitations of estimations

    The massive internal displacement of more than 14 million people in Sudan complicates the estimation of death tolls, as shifting populations make establishing baselines nearly impossible.

    Moreover, there is typically a dearth of official information collected and released during conflicts.

    So establishing a concrete estimate of the true impact of armed conflict often comes after the cessation of hostilities, when expert teams are able to conduct field studies.

    Even then, estimates will require assumptions about direct deaths, indirect-to-direct death ratio and the quality of existing data.

    But as scholars working at the intersection of public health and human rights, we believe such work, however imperfect, is necessary for the documentation of conflict – and its future prevention. And while there are many current global conflicts that require our urgent attention, the conflict in Sudan must not be lost in the mix.

    _Editor’s note: Israa Hassan, a physical medicine and rehabilitation resident at Texas Rehabilitation Hospital-Fort Worth and advocacy director at the Sudanese American Physicians Association, contributed to this article.

    Rohini J Haar receives funding from FCDO.

    Blake Erhardt-Ohren, Debarati Guha Sapir, Khidir Dalouk, and Sarah Elizabeth Scales do not work for, consult, own shares in or receive funding from any company or organization that would benefit from this article, and have disclosed no relevant affiliations beyond their academic appointment.

    ref. Sudan’s civil war has left at least 62,000 dead by our estimate − but the true figure could be far higher – https://theconversation.com/sudans-civil-war-has-left-at-least-62-000-dead-by-our-estimate-but-the-true-figure-could-be-far-higher-242073

    MIL OSI – Global Reports

  • MIL-OSI Global: World Update: how Israel’s relations with the UN hit rock bottom

    Source: The Conversation – UK – By Jonathan Este, Senior International Affairs Editor, Associate Editor

    With the clock ticking down to November 5 and what just about everyone agrees is the most consequential US presidential election in living memory, various of the Biden administration’s top brass have jetted out to the Middle East for one last try to get a deal over the line.

    The most likely area where progress could be made is the conflict in Lebanon between Israel and Hezbollah. The militant group announced the appointment of a new general secretary on October 29. Naim Qassem is, as the BBC puts it, “one of the few senior Hezbollah leaders who remains alive after Israel killed most of the group’s leadership in a series of attacks”. He is reportedly making noises about possible change in Hezbollah policy that would separate any negotiations over the conflict in Lebanon with any talks over Gaza.

    If true, it’s a major shift from the policy of recently assassinated leader Hassan Nasrallah, which previously indelibly linked a ceasefire in Gaza with the cessation of Lebanon’s rocket attacks on northern Israel. Full details of the deal remain under wraps, but a draft was leaked to Israel’s state broadcaster Kan.

    Post on X by Kann reporter, Suleiman Maswadeh, with details of a proposed Middle East peace deal.
    X

    For Israel’s part, Prime Minister Benjamin Netanyahu has said the initial phase of Israel’s operation inside Lebanon is drawing to a close. As for what comes next, the New York Times reported on October 28 that Netanyahu is “waiting to see who will succeed President Biden before committing to a diplomatic trajectory”.

    The diplomatic trajectory has been made more complicated of late by a big spat between Israel and the UN. The two have had a fractious relationship since the very start. But under the Netanyahu government, things have steadily deteriorated to the stage that Israel actually barred UN secretary general António Guterres from entering the country at the beginning of October.

    This week Israel’s parliament, the Knesset, passed a new law banning the UN relief and works agency (Unrwa) from operating on any territory it controls. Unrwa was set up after the war of 1948 to help displaced Arabs and has since morphed into what an independent review this year said was an “indispensable lifeline” for civilians in Gaza and the West Bank.

    The trouble is that the reason the independent review was reporting at all was that Israel was alleging Unrwa staff had taken part in the October 7 massacres alongside Hamas. Unwra subsequently fired nine staff members. But Israel’s contention that Unrwa is a “rotten tree entirely infected with terrorist operatives” remains unproved.

    Lisa Strömbom of Sweden’s Lund University, who has been following the conflict for many years, has traced the deterioration of relations between Israel and Unrwa over several decades. She now believes that Israel’s ban will make it nigh on impossible for Unrwa to fulfil its mission in Gaza. This can only make things worse for a civilian population in Gaza which is already trying to survive in the most difficult circumstances possible.




    Read more:
    Israel’s relations with the UN hit a new low with Unrwa ban


    The Netanyahu government’s decision to ban Unrwa has been roundly condemned on all sides. Some voices have even called for Israel’s membership of the UN to be suspended. That’s a complicated issue, writes Aidan Hehir, who has published widely on conflict resolution and treaty making.

    For a start, it would need to get past the UN security council which means being subject to a veto from any one of the five permanent members (P5). We published an article on this issue some years ago with the help of UN expert Emma McClean, which looked at the issues which had prompted members of the P5 to wield their vetos. It found that Israel-Palestine was hands-down the most common issue that led to a veto – and all those vetoes had been instigated by the US.

    UN security council vetoes.
    UN security council



    Read more:
    Hard Evidence: who uses veto in the UN Security Council most often – and for what?


    So suspending Israel from the UN would appear to be a non-starter. But Hehir tells the story of the way the UN managed to circumvent the P5 and suspend South Africa in 1974 over apartheid. Having failed to get the suspension past the security council after the UK and France vetoed the move, the credentials committee of the general assembly simply refused to renew South Africa’s credentials. It remained suspended for two decades until the end of apartheid in 1994.




    Read more:
    Gaza: can the UN suspend Israel over its treatment of Palestinians? It’s complicated, but yes


    Meanwhile Israel’s assault on Gaza continues and the death toll continues to mount. Israeli Defense Forces (IDF), supported by airstrikes, continue to bombard what the IDF says are Hamas positions in the towns of Beit Lahia and Jabalia but which the Gaza health ministry say are residential buildings sheltering hundreds of civilians. On October 29, the health ministry said at least 93 people, including 25 children, were killed by an Israeli airstrike.


    Now, more than ever, it’s vital to be informed about the important issues affecting global stability. Sign up to receive our weekly World Update newsletter. Every Thursday we’ll you expert analysis of the big stories making international headlines.


    Much of the population of the north of Gaza has been evacuated south of what is known as the Netzarim corridor. Israel’s Haaretz newspaper claims that it’s part of an operation known as the “generals’ plan”, which calls for the north to be cleared of civilian residents and locked down as a military zone. This is presented as a national security measure, but Leonie Fleischmann reports that there are those who believe the military operation will be followed by an influx of Israeli settlers.

    Fleischmann points to a conference held on the Israeli side of the border with north Gaza, attended by members of Netanyahu’s Likud party as well as by several government ministers, which actively promoted the idea of settling north Gaza. Memories and historical legend mingle with ideology that holds Gaza had a Jewish population from biblical times through to 1929, when an Arab revolt killed 133 Jewish people living there and drove the rest out.

    The prospect of a land grab is clearly exercising minds at the UK foreign office. UK ambassador to the United Nations, Barbara Woodward said on October 29: “We reiterate that northern Gaza must not be cut off from the south. Palestinian civilians, including those evacuated from northern Gaza must be permitted to return. There must be no forcible transfer of Gazans from or within Gaza, nor any reduction in the territory of the Gaza Strip.”




    Read more:
    Israel’s ‘generals’ plan’ to clear Palestinians from north of Gaza could pave the way for settlers to return


    All eyes on Washington

    It’s highly unlikely that we’ll know by this time next week who has prevailed in the US presidential election. But the Middle East will be one of the first big ticket items on the Resolute desk.

    The issue has already proved to be a tricky one for Kamala Harris. Her support base is deeply divided on the issue, with large numbers of Democrats – particularly young people, as well as Muslims and black voters – unsettled by her perceived part in the Biden administration’s “steadfast” support for Israel over the past four years.

    It’s hard to tell whether these voters consider that the people of Gaza would fare any better under a Trump White House. But Natasha Lindstaedt and Faten Ghosn believe that Netanyahu’s continuing aggression in Gaza may well play out in the Republican contender’s favour.




    Read more:
    How the Middle East conflict could influence the US election – and why Arab Americans in swing states might vote for Trump


    Meanwhile, to guide us through how the two candidates are likely to approach the big foreign policy issues, we can turn to Garret Martin of the Transatlantic Policy Center at the American University in Washington.




    Read more:
    On foreign policy, Trump opts for disruption and Harris for engagement − but they share some of the same concerns


    World Update is available as a weekly email newsletter. Click here to get our updates directly in your inbox.


    ref. World Update: how Israel’s relations with the UN hit rock bottom – https://theconversation.com/world-update-how-israels-relations-with-the-un-hit-rock-bottom-242632

    MIL OSI – Global Reports

  • MIL-OSI USA: Through Medical Education UConn Professor Helps Maintain Normalcy in War-Torn Ukraine

    Source: US State of Connecticut

    This fall in Lviv, Ukraine for three days over 250 Ukrainian health care professionals gathered together for a medical conference to take a master class in radiology and participate in hands-on training workshops.

    The conference held Sept. 25-27 was co-chaired by UConn School of Medicine professor and chair of Diagnostic Imaging & Therapeutics Dr. Leo Wolansky who also delivered several lectures.

    Dr. Leo Wolansky.

    “I am so impressed with the spirit of the Ukrainian people. They insist on keeping their lives normal. Hopefully, our resuming our in-person conference series, now in its twenty-eighth year, contributed a little bit to that normalcy,” says Wolansky.

    The three-day conference was also co-chaired by Dr.  Tetyana Yalynska. It was organized by Friends Of Radiology in Ukraine (FORUkraine), the Ukrainian Medical Association of North America (UMANA), and the Association of Radiologists of Ukraine (ARU), and was hosted by the First Territorial Medical Association of Lviv. Additional sponsorship came from Rotary, St. Paraskeva’s, Azaris, Ulrich, and Guerbet.

    This year’s event carries forward the FORUkraine conference series founded by Wolansky in 1996 at the Lviv Medical University (the Lviv Medical Institute). Since its inception, the conference series has been innovative using state-of-the-art education as a tool to promote the Ukrainian language. The purpose of the program is to teach Ukrainian-language based, state-of-the-art Diagnostic Imaging (Radiology).

    The long-standing in-person conference was interrupted for the last five years due to the COVID-19 crisis and followed by the war in Ukraine. Starting in 1998, Dr. Yuriy Ivaniv, head of post-graduate imaging education for the Lviv Oblast co-chaired the conferences with Wolansky and renamed it “Practical Questions in Contemporary Clinical Imaging.”  The conference series continued every one or two years from then till 2019 when the program joined forces with the ARU and the American Society of Neuroradiology and held a combined conference with the Ukrainian Congress of Radiology, with ARU President Yalynska co-chairing with Wolansky. This took place in Kyiv and Irpin, the only time the program deviated from its home in Lviv.

    Despite the war, several international imaging experts attended the event in-person in Ukraine to lecture. One of these was Dr. John (Ivan) Kachura, an Interventional Radiologist and professor of Medical Imaging at University of Toronto, who stated that he was impressed by the knowledge of Ukraine’s radiologists, but especially by their tremendous interest in the presented material. Also onsite was Dr. Andrew Dobrotwir, consultant radiologist from Melbourne, Australia, who lectured and also ran a hands-on workshop on Point Of Care Ultrasound (POCUS). Dobrotwir was accompanied by his sister-in-law, Teresa Lachowicz, who spoke at the conference about humanitarian work that she and Dobrotwir were carrying out including donations of POCUS equipment.

    Several international lecturers also virtually presented for the conference including Laura Oleaga, Amy Juliano, Diana Kaya, Andrew Loginsky, and Ivan Wolansky. In addition, virtual presentations came from UConn’s Jill Wruble, Sarah LaPierre, and Racquel Helsing. Several Ukrainian physicians also presented at the conference, including Nataliya Deresh, Igor Ivaniv, and others.

    Nataliya Motrynets, medical director of the host institution, presented about the hospital’s accomplishments, and also gave the faculty a tour of the facilities where many soldiers who have lost limbs in the war are being rehabilitated with cutting-edge prostheses.

    For Wolansky, a semiprofessional musician, a highlight of his trip to Ukraine was when he and his wife, Maria, met Dudaryk, Lviv’s internationally renowned boys choir, at Mass on the last day of the trip. The children’s choir had performed the refrain for Wolansky’s recent music video, helping draw attention to the plight of Ukraine’s children.

    Wolansky added, “Despite air alerts driving conference participants into the bomb shelter on two occasions, these brave Ukrainians insist that normal life must go on!”

    MIL OSI USA News

  • MIL-Evening Report: What should I do to prepare for a monologue performance?

    Source: The Conversation (Au and NZ) – By Natasha Beaumont, Lecturer in Creative Arts, School of Education and Social Work, University of Sydney

    Frame Stock Footage/Shutterstock

    Monologue performance is a technically demanding but deeply rewarding form of theatre. Monologues are the purest form of storytelling an actor can engage in.

    Before I was a drama teacher and researcher, I was an actor on stage and television in Australia and in the United Kingdom.

    As an actor, you are always having to prepare monologues as audition pieces. Here are some principles and techniques to help you with this process, to allow you to draw in your audience and strengthen your artistic expression.

    Choosing your monologue

    Successful characterisation in monologue acting depends on “casting” yourself well. This means choosing a text and a character that resonate with your own persona and emotional range.

    Even if you decide you want to play someone with a completely different age, gender and life experience to your own, there should be something about this character that speaks to you: maybe it’s their sense of vulnerability, their love of life, their rage at unfair circumstances. The more you can relate to some specific aspect of this person, the easier it will be to access the emotional range to play them.

    If you decide on a character from a well-known play, make sure you have an understanding of the whole text the piece comes from.

    Finding the personality

    As an actor, you should have a good grasp of your character’s personality and attitudes to life.

    Look for clues in the monologue or the overall play that tell you something about this person’s inner psychology. Do they always agree with everyone, or are they always complaining? How do they talk about themselves, how do they talk about other people?

    In a well-written play, dialogue is always filled with signals like these that actors rely on when creating characters.

    Get clues about your character’s personality from the script.
    Cynthia Smith/Unsplash

    Another useful approach is to develop a detailed backstory for the role you are playing. Performers often use journalling or visualisation to deepen their emotional connection with the person they are depicting.

    Taking time to imagine these key “memories” can provide an emotional anchor when you want to access different parts of their personality. The audience will never know these choices you have made, but you will carry them within you, and they can add depth and dimension to your portrayal.

    Making the character physical

    Along with analysing your character’s psychology and motivations, spend time working on their physicality.

    How does this person move through the world? Are they a daydreaming wanderer, or a short sharp stepper who is always in a hurry? Do they close themselves off from the world with hunched shoulders, or do they stand tall and project themselves outward?

    These qualities might change throughout the monologue as your character moves through different thoughts and memories.

    Making stage direction choices for a monologue can be one of the most challenging things to get right. Simple things such as walking downstage to talk directly to your audience, or sitting down at a particular moment, can add effective dynamics to your performance. But any choices you make must come from an inner impulse within your character. Movement needs to be motivated by some kind of shift in their thoughts.

    Breaking down the monologue

    To identify these shifts, break down your script into key “beats”. These are the moments in a text where your character starts talking about something new. You can use these to create shifts in movement, tone and pace.

    Incorporating different beats into your piece is vital for keeping your audience’s interest. Every monologue should take the audience on a journey through a character’s inner life. Ensuring this journey includes some surprises or effective use of dramatic tension will help make your piece work as a solo performance.

    Sit down with the script and a pencil to find the ‘beats’ of the monologue.
    Media_Photos/Shutterstock

    Sit down with a pencil and mark down any point in the script where you think the character starts thinking or talking about something new. Once you have all these internal shifts marked out, decide if any of these could be played with a contrasting emotional tone and pace to create dramatic effect.

    Who are you talking to?

    Performing a piece on your own can be daunting as there are no other characters to respond to or generate reactions from. Understanding who your character is speaking to during the monologue means you can use your audience as an additional “actor”. Are they an ally or an enemy? Or are these private thoughts, with the audience as a witness to your inner mind?

    Clarifying this relationship can help you make clearer choices in how you deliver your lines.

    Give yourself time

    There are many creative decisions to be made when preparing a monologue performance.

    Make sure to give yourself enough time to make these decisions and to learn your lines by heart.

    Think about the physicality of the character you have chosen.
    foto-lite/Shutterstock

    Experiment with lots of different choices when you are starting out and rehearse your piece as often as possible. This will help reduce nerves when it comes to your final performance as it’s difficult to focus on acting when your mind is racing trying to remember what to say next.

    Once the hard work of preparation, experimentation and creative expression is done, there is no better feeling than nailing a solo performance!

    Natasha Beaumont does not work for, consult, own shares in or receive funding from any company or organisation that would benefit from this article, and has disclosed no relevant affiliations beyond their academic appointment.

    ref. What should I do to prepare for a monologue performance? – https://theconversation.com/what-should-i-do-to-prepare-for-a-monologue-performance-238778

    MIL OSI AnalysisEveningReport.nz

  • MIL-Evening Report: The ‘big 4’ accounting firms often consult for the same clients they audit. Should that be allowed?

    Source: The Conversation (Au and NZ) – By Helen Spiropoulos, Associate Professor, University of Technology Sydney

    Public trust in the auditing profession is under intense pressure. A series of high-profile scandals, both in Australia and overseas, has severely damaged its reputation.

    This week, Australia’s corporate watchdog – the Australian Securities and Investments Commission (ASIC) – put the entire sector on notice.

    In a letter to auditors on Wednesday, ASIC announced it would soon commence a new data-driven surveillance of auditor independence and conflicts of interest. Put simply, any practices that could compromise the integrity of auditing work.

    The move comes amid longstanding calls for stronger regulation. Some have gone as far as to call for auditors – particularly the “big four” – to be banned from offering consulting services to their audit customers. Why? Fears it helps companies unethically game the system.

    But our recent research, which specifically examines chief executive pay, offers an alternative perspective and suggests we should tread carefully.




    Read more:
    A year after the PwC scandal, the furore is gone – as well as any real appetite for structural change


    Objectivity and independence

    The “big four” – PricewaterhouseCoopers (PwC), Ernst & Young (EY), KPMG and Deloitte – are the world’s largest professional services firms. They offer services in auditing, consulting, tax and advisory services.

    Known for their extensive resources and global reach, these firms serve major clients, including many publicly listed companies and governments.

    However, some have raised concerns about potential conflicts of interest that may arise when these firms provide both consulting and auditing to the same client.

    Auditing is the process of examining a company’s financial statements and processes to ensure both accuracy and compliance with accounting standards.

    Conducted by external auditors, it’s meant to give investors, regulators, and the public confidence that a company’s financial picture is accurate and trustworthy.

    The key worry is that offering both services risks compromising an auditor’s objectivity and independence.

    Auditors may be incentivised to shy away from scrutinising their clients too closely, if it helps preserve lucrative consulting contracts.

    How much money should the boss make?

    Professional services firms, including the big four, are often engaged as external consultants to help decide on “executive compensation” – how much a company’s chief executive should be paid.

    Chief executive pay is highly contentious. They can earn staggering amounts of money, which can sometimes appear disconnected from how well a company is actually performing and what’s in its shareholders’ best interests.

    Large companies often outsource decisions about how much to pay chief executives.
    GaudiLab/Shutterstock

    Compensation consultants are hired to help structure these pay packages, ideally by setting up performance targets that align chief executives’ incentives with shareholder value.

    The idea is that if you don’t meet a certain goal as the boss, you should miss out on being paid for it.

    But these consultants can also be a part of the problem. As chief executives can influence whether a particular consultant is hired or retained, consultants might design favourable contracts to increase their chances of getting hired again.

    How? By setting up targets that are easy to hit, or vague enough to avoid true accountability.

    Such accountability in executive compensation is extremely important. How much those at the top get paid should reflect the quality of their decisions.

    Without proper oversight, pay structures risk incentivising quick wins instead of long-term growth, which could potentially harm investors, employees and the company’s future.

    To solve this problem, you need transparent performance metrics. This makes it easier for shareholders to see whether chief executives are truly earning their pay.

    When executive compensation consultants do their job well, such transparency gets built in. So how does the big four score?

    What we found

    Our study, published in the Australian Journal of Management, analysed chief executives’ compensation structures in a sample drawn from the 500 largest companies listed on the Australian Securities Exchange (ASX), between 2005 and 2019.

    We found that the big four, when engaged as compensation consultants, appeared to uphold more rigorous standards than their smaller counterparts.

    For example, big four firms were more likely to recommend including performance measures like “relative total shareholder return”, which takes the performance of a company’s competitors into account.

    This can reduce the likelihood of “pay for luck” – paying a chief executive extra when a company performs well simply due to market-wide factors, such as movements in commodity prices or currency exchange rates.

    Non-big four consultants, on the other hand, showed a tendency towards less clearly defined targets, which can open the door to less accountability.

    Compensation consultants should set targets for chief executives that genuinely reflect good performance.
    Owlie Productions/Shutterstock

    What’s behind this effect?

    One possible explanation for our findings is that the big four’s multi-service approach gives them less reliance on securing repeat business from any single client.

    With consulting, tax, audit and advisory services across various industries, these firms aren’t as dependent on individual clients, which can give them greater freedom to recommend compensation packages that may not always align with a chief executive’s preferences.

    It has been argued, including by former chairman of the Australian Competition and Consumer Commission Graeme Samuel, that the big four’s consulting services pose potential conflicts that could compromise their audit duties.

    The same could be said for other advisory services provided by these firms.

    However, our findings offer evidence that when it comes to executive compensation, the big four’s reputation and expertise may actually discourage practices that obscure performance metrics or result in excessive chief executive pay.

    Any reforms should tread carefully

    The auditing sector will be watching the outcomes of ASIC’s forthcoming “crackdown” closely. The case for stricter oversight is strong.

    But we should be careful not to lose the nuance of this issue. In some cases, the big four’s multi-service approach may actually elevate governance standards rather than erode them.

    In a market dominated by these firms, the consequences of their exit from consulting services could extend beyond audit independence.

    Ironically, forcing these firms out of consulting could make auditing their primary revenue source from many clients, creating the very dependence regulators aim to avoid.

    Are we ready to face the unintended effects of limiting these firms’ roles? If our research is any indication, the answer is not so clear-cut.

    As an undergraduate student, Helen Spiropoulos did two internships at Deloitte in the areas of Audit and then Consulting (Strategy and Operations).

    Rebecca L. Bachmann does not work for, consult, own shares in or receive funding from any company or organisation that would benefit from this article, and has disclosed no relevant affiliations beyond their academic appointment.

    ref. The ‘big 4’ accounting firms often consult for the same clients they audit. Should that be allowed? – https://theconversation.com/the-big-4-accounting-firms-often-consult-for-the-same-clients-they-audit-should-that-be-allowed-242588

    MIL OSI AnalysisEveningReport.nz

  • MIL-Evening Report: In Norway, students get grades for their behaviour – could this work in Australia?

    Source: The Conversation (Au and NZ) – By Stephen Dobson, Professor and Dean of Education and the Arts, CQUniversity Australia

    Student behaviour is one of the biggest issues facing Australian schools. A survey of Queensland teachers earlier this year found “managing student behaviour” was the main thing taking their time away from teaching.

    Along with students talking out of turn, using their phones or not paying attention, there are regular reports of students being violent and abusive towards teachers. Australian classrooms are rated among the “least favourable” for discipline in the OECD.

    Amid a push to include more classroom management training for teachers, what other approaches could we look at to improve behaviour?

    What happens in Norway?

    For several decades Norwegian school children have been assessed twice a year on their sense of personal order (being punctual, well-prepared and following up on homework) and social behaviour (showing care and respect for others).

    In some schools this might involve following rules against throwing snowballs, eating in class or leaving school grounds.

    Until Year 8, students receive comments and then they also get a grade (good, quite good or not so good).

    Teachers in all subjects report to the child’s home base teacher who calculates an average, noting any poor examples of poor personal order and social behaviour. The overall report is shared with the student and parents receive a copy.

    The goal, as specified in Norway’s Education Act, is to ensure a good and safe school environment and “social learning”. This means learning to behave around others through observing, modelling and imitating the behaviours of others.

    This is on top of learning knowledge and skills.

    Norwegian students can be graded on whether they follow rules about snowball throwing.
    Maria Sbytova/Shutterstock

    Does it work?

    Norwegian society takes these grades seriously. It has been part of the Norwegian schooling system since 1939.

    Research on teachers and students describe it as a valued tool for dealing with students who disrupt the learning environment in the classroom.

    Even when young adults apply for jobs after university or vocational study, employers can be interested in the grade received for order and behaviour at school. Students and their teachers are aware it can indicate trustworthiness and employability.

    A not uncommon story repeated by Norwegian parents to their teenage children is “if you have a record of behaving poorly or arriving late at school it doesn’t bode well whether you want to work on a construction site, in an office or on a hospital ward”.

    There are Norwegian critics of this approach. Some researchers argue behaviour grades can sometimes say more about who are the “teachers’ favourites”.

    But despite some limited trials to refine Norway’s behaviour grading, there are currently no plans to remove it.

    What about Australia?

    There is some precedence for reporting on behaviour in Australia.

    For example, Queensland public schools report about effort and behaviour against a five-point scale: excellent, very good, satisfactory, needs attention and unacceptable.

    But assessment criteria and evidence for the reporting of student effort and behaviour seems to be a more subjective appraisal than reporting against other standards in the curriculum.

    Some Australian schools already report on aspects of student behaviour.
    Monkey Business Images/ Shutterstock

    School is about more than maths and reading

    Schools can teach students more than academic knowledge or vocational skills.

    And while addressing behaviour in schools is complex (and will not be solved by any single thing), reporting on behaviour could provide a regular opportunity for Australian teachers, schools and parents to reflect on how a students is progressing.

    Grading students could make students more accountable for how they interact with their peers and their teachers.

    It could also help build their understanding of what is acceptable, not just in the classroom but in the community more broadly. For example, if there are specific rules about how you speak to others, whether you are safe in the playground and respectful in the classroom.

    This type of social learning is important, because it can help teach students to be inclusive and responsible towards others. It can also help to create a safer school environment for all students and staff.

    At the moment, there is a general requirement in the Australian Curriculum to teach students social and emotional skills across all subjects.

    But it is up to state and territory education authorities to work out if and how students are assessed about this. This includes any reasonable adjustments for students with disability or other special needs.

    The authors do not work for, consult, own shares in or receive funding from any company or organisation that would benefit from this article, and have disclosed no relevant affiliations beyond their academic appointment.

    ref. In Norway, students get grades for their behaviour – could this work in Australia? – https://theconversation.com/in-norway-students-get-grades-for-their-behaviour-could-this-work-in-australia-239384

    MIL OSI AnalysisEveningReport.nz

  • MIL-Evening Report: From eye exams to blood tests and surgery: how doctors use light to diagnose disease

    Source: The Conversation (Au and NZ) – By Matthew Griffith, Associate Professor and ARC Future Fellow and Director, UniSA Microscopy and Microanalysis Facilities, University of South Australia

    megaflopp/Shutterstock

    This is the next article in our ‘Light and health’ series, where we look at how light affects our physical and mental health in sometimes surprising ways. Read other articles in the series.


    You’re not feeling well. You’ve had a pounding headache all week, dizzy spells and have vomited up your past few meals.

    You visit your GP to get some answers and sit while they shine a light in your eyes, order a blood test and request some medical imaging.

    Everything your GP just did relies on light. These are just some of the optical technologies that have had an enormous impact in how we diagnose disease.

    1. On-the-spot tests

    Point-of-care diagnostics allow doctors to test patients on the spot and get answers in minutes, rather than sending samples to a lab for analysis.

    The “flashlight” your GP uses to view the inside of your eye (known as an ophthalmoscope) is a great example. This allows doctors to detect abnormal blood flow in the eye, deformations of the cornea (the outermost clear layer of the eye), or swollen optical discs (a round section at the back of the eye where the nerve link to the brain begins). Swollen discs are a sign of elevated pressure inside your head (or in the worst case, a brain tumour) that could be causing your headaches.

    The invention of lasers and LEDs has enabled many other miniaturised technologies to be provided at the bedside or clinic rather than in the lab.

    Pulse oximetry is a famous example, where a clip attached to your finger reports how well your blood is oxygenated. It does this by measuring the different responses of oxygenated and de-oxygenated blood to different colours of light.

    Pulse oximetry is used at hospitals (and sometimes at home) to monitor your respiratory and heart health. In hospitals, it is also a valuable tool for detecting heart defects in babies.

    See that clip on the patient’s finger? That’s a pulse oximeter, which relies on light to monitor respiratory and heart health.
    CGN089/Shutterstock

    2. Looking at molecules

    Now, back to that blood test. Analysing a small amount of your blood can diagnose many different diseases.

    A machine called an automated “full blood count analyser” tests for general markers of your health. This machine directs focused beams of light through blood samples held in small glass tubes. It counts the number of blood cells, determines their specific type, and reports the level of haemoglobin (the protein in red blood cells that distributes oxygen around your body). In minutes, this machine can provide a snapshot of your overall health.

    For more specific disease markers, blood serum is separated from the heavier cells by spinning in a rotating instrument called a centrifuge. The serum is then exposed to special chemical stains and enzyme assays that change colour depending on whether specific molecules, which may be the sign of a disease, are present.

    These colour changes can’t be detected with the naked eye. However, a light beam from an instrument called a spectrometer can detect tiny amounts of these substances in the blood and determine if the biomarkers for diseases are present, and at what levels.

    Light shines through the blood sample and tells us whether biomarkers for disease are present.
    angellodeco/Shutterstock

    3. Medical imaging

    Let’s re-visit those medical images your GP ordered. The development of fibre-optic technology, made famous for transforming high-speed digital communications (such as the NBN), allows light to get inside the body. The result? High-resolution optical imaging.

    A common example is an endoscope, where fibres with a tiny camera on the end are inserted into the body’s natural openings (such as your mouth or anus) to examine your gut or respiratory tracts.

    Surgeons can insert the same technology through tiny cuts to view the inside of the body on a video screen during laparoscopic surgery (also known as keyhole surgery) to diagnose and treat disease.

    Doctors can insert this flexible fibre-optic tube with a camera on the end into your body.
    Eduard Valentinov/Shutterstock

    How about the future?

    Progress in nanotechnology and a better understanding of the interactions of light with our tissues are leading to new light-based tools to help diagnose disease. These include:

    • nanomaterials (materials on an extremely small scale, many thousands of times smaller than the width of a human hair). These are being used in next-generation sensors and new diagnostic tests

    • wearable optical biosensors the size of your fingernail can be included in devices such as watches, contact lenses or finger wraps. These devices allow non-invasive measurements of sweat, tears and saliva, in real time

    • AI tools to analyse how blood serum scatters infrared light. This has allowed researchers to build a comprehensive database of scatter patterns to detect any cancer

    • a type of non-invasive imaging called optical coherence tomography for more detailed imaging of the eye, heart and skin

    • fibre optic technology to deliver a tiny microscope into the body on the tip of a needle.

    So the next time you’re at the GP and they perform (or order) some tests, chances are that at least one of those tests depend on light to help diagnose disease.

    Matthew Griffith receives funding from the National Health and Medical Research Council and the Australian Research council.

    ref. From eye exams to blood tests and surgery: how doctors use light to diagnose disease – https://theconversation.com/from-eye-exams-to-blood-tests-and-surgery-how-doctors-use-light-to-diagnose-disease-231379

    MIL OSI AnalysisEveningReport.nz

  • MIL-Evening Report: New data on violence against LGBTQ+ people makes ‘grim reading’ – and undermines NZ’s inclusive reputation

    Source: The Conversation (Au and NZ) – By Alexander Plum, Senior Research Fellow, Auckland University of Technology

    New Zealand is ranked the tenth most inclusive society by international legal standards, with a reputation for being forward-thinking and progressive – especially when it comes to the rights of sexually and gender-diverse individuals.

    But recent high-profile hate crimes against the LGBTQ+ community suggest we may not be as progressive as our global reputation suggests.

    The painting over of rainbow pedestrian crossings in Gisborne and Auckland might seem like comparatively minor crimes. But they highlight the insidious – and increasingly overt – nature of prejudice against the rainbow community.

    A major concern for members of this community is how easily this kind of prejudice spills over into criminal acts against them. And there are indications of a concerning trend. The number of reported hate crimes against transgender people rose by 42% between 2022 and 2023.

    This is backed by overseas research. According to a study from the United States, gay/lesbian and bisexual individuals are significantly more likely to be victims of violence than heterosexual men and women.

    But how do rates of violence and crime faced by LGBTQ+ individuals here compare to the general population in New Zealand? For the first time, our new research sheds light on crime victimisation rates among the LGBTQ+ population in New Zealand. It’s grim reading.

    High rates of crime victimisation

    Our research used data from the New Zealand Crime and Victims Survey (NZCVS).

    The NZCVS surveyed approximately 32,000 New Zealanders between 2018 to 2022. Participants were asked about their experiences with crime. The survey collected information on reported and non-reported offences, and asked the participants about their socio-demographic characteristics, including sexual orientation and gender identity.

    LGBTQ+ individuals include those whose sexual orientation is
    gay/lesbian, bisexual, or other, or when being gender diverse or when gender identity and biological sex differ (also called transgender).

    We found LGBTQ+ individuals were much more likely to be victims of crime than non-LGBTQ+ individuals.

    Almost half of LGBTQ+ respondents (46%) reported being a victim of at least one crime in the previous year, compared to a third of non-LGBTQ+ people (31%).




    Read more:
    NZ’s hate speech proposals need more detail and wider debate before they become law


    Members of the LGBTQ+ community were also much more likely to be a victim of crime more than once. According to the survey, approximately 22% of LGBTQ+ individuals experienced more than one victimisation in the previous year, compared with 11% of non-LGBTQ+ individuals.

    Two groups stood out in particular: bisexual individuals and transgender/gender-diverse people.

    One potential explanation for the crime rates against LGBTQ+ people is that they have higher-than-average risk factors that are unrelated to their sexual orientation or gender identity. For example, they are younger and have lower incomes on average.

    But our research refutes this explanation. Even after accounting for these other risk factors, the crime victimisation rates among LGBTQ+ individuals were much higher than among non-LGBTQ+ individuals.

    Motivating factors

    The NZCVS also collected information on the perceived motivation behind the crime. Response options included sexual orientation, sex or general discrimination.

    We found LGBTQ+ individuals were more likely to say the perceived reason for crime was their sexual orientation or their sex compared to non-LGBTQ+ individuals.

    The consequences of these offences were also more severe for LGBTQ+ individuals.

    They were more likely to suffer from physical injuries or need time off work. They were also more likely to feel less noticeable effects of the violence: lower life satisfaction and a greater sense of feeling unsafe.

    Living up to NZ’s inclusive reputation

    In the long term, understanding how victimisation affects LGBTQ+ individuals can help shape policies that are better tailored to prevent crime and support victims. This includes building greater awareness and knowledge in the sexual and family violence sectors to prevent and support affected rainbow communities.

    But until that happens, crime victimisation continues to disproportionately affect LGBTQ+ individuals. New Zealand needs to do something to close that gap.

    Our research highlights a serious gap between how New Zealand is perceived on the global stage (safe and inclusive), and the reality of life for our LGBTQ+ community (increasingly unsafe and threatened by intolerance).

    New Zealand’s laws must ensure crimes against people based on their sexual orientation and gender identity will not be tolerated. The ongoing review of the Human Rights Act is a step in the right direction but more needs to be done to explicitly protect trans, non-binary and intersex people against discrimination.


    The authors want to thank Tabby Besley for her feedback. Tabby is the managing director at InsideOut, which provides resources, workshops, consulting, advocacy and support for anything concerning rainbow communities


    Alexander Plum receives funding from the Ministry of Justice.

    Lee Zhuge receives funding from The Ministry of Justice of New Zealand.

    ref. New data on violence against LGBTQ+ people makes ‘grim reading’ – and undermines NZ’s inclusive reputation – https://theconversation.com/new-data-on-violence-against-lgbtq-people-makes-grim-reading-and-undermines-nzs-inclusive-reputation-239706

    MIL OSI AnalysisEveningReport.nz

  • MIL-Evening Report: Explainer: a short history of the Electoral College and how it subverts the will of voters

    Source: The Conversation (Au and NZ) – By John Hart, Emeritus Faculty, Australian National University

    For a fleeting moment in early October, it looked like the US presidential electoral system might become an issue in this year’s election. The Democratic vice presidential candidate, Tim Walz, told two audiences that the Electoral College should be abolished and replaced by a direct national popular vote.

    Walz was shut down quickly by Kamala Harris’ campaign with a brief statement that abolishing the Electoral College is not its official position. Walz duly walked back his comments and the story had a shelf-life of fewer than 24 hours.

    But the Electoral College issue may well come back to haunt the Harris campaign should this year’s election produce yet another “runner-up” president – when the loser of the popular vote wins the electoral vote and therefore the election.

    If the race is as close as most polls are indicating, this is a possible outcome. And Republican former President Donald Trump is more likely than Harris to be the beneficiary of this archaic, undemocratic voting system.

    How the Electoral College works

    There is a two-stage, indirect election for the president under the Electoral College system.

    First, there is the popular vote in each of 50 states and District of Columbia on November 5 to choose “electors”, who formally cast the “electoral vote” on December 17 in what is known as the “Electoral College”.

    It is the electoral vote that determines the president, not the popular vote.

    To make things even more complicated, each state is awarded electoral votes based not on its population, but on its representation in the US Congress.

    Each state has at least one member of the House of Representatives and two members of the Senate, meaning every state has at least three electoral votes regardless of its population size.

    There are 538 votes in the Electoral College, and an absolute majority of those – 270 or more – is needed to win. The Constitution also contains a complex and highly undemocratic contingency procedure should no candidate win an Electoral College majority. The choice of president would then be decided by the House of Representatives with each state delegation having just one vote.

    Sample presidential ballot from Arlington County in the state of Virginia showing that voters will be selecting electors, not the candidate directly.
    Arlington County Electoral Board

    The origins of the Electoral College

    It is not surprising the Electoral College is an undemocratic institution – it was deliberately designed to be. The method of electing the president was an expression of a very conservative philosophy of government embodied by most of the framers of the Constitution when they met in Philadelphia in 1787.

    The framers had strong views the presidency should be an office above politics. They also felt the choice should be made by those with knowledge, experience and understanding of government and statecraft.

    As such, the framers objected to a popular vote for the president, because they feared it would lead to what one of the founding fathers, Alexander Hamilton, called “tumult and disorder”. The framers were vehemently opposed to direct democracy, preferring instead what they called a “republic”.

    Their solution was to allow the state legislatures to determine how the electors from each state should be chosen. In the beginning, most states’ legislatures chose the electors to decide who was president – not the people.

    The Electoral College structure – and its philosophical underpinnings – were then locked into the Constitution and purposely designed to exclude the people from the process.

    It has also been argued that race and slavery were integral to its design. By piggy-backing on the already-agreed compromise over representation in Congress and the counting of slaves as “three-fifths of all other persons”, the framers of the Constitution handed the major slave-holding states far more clout not only in Congress, but in the selection of the president, as well.

    In the longer term, the framers weren’t entirely successful in their efforts because two major political developments in the early 19th century forced some adaptation to the model.

    As the American frontier expanded and political parties were developed, people began demanding a greater role in American democracy. This put pressure on state legislatures to cede their power to select electors and allow popular voting for the Electoral College instead.

    By the mid-19th century, the Electoral College was operating in much the same way as it does today.

    Surprisingly, this required no constitutional amendment because the wording of the Constitution gave the states the flexibility to respond to the demand for popular voting:

    Each State shall appoint, in such manner as the legislature thereof may direct, a number of Electors…

    But that didn’t change the fact that it was the “electors” who would still choose the president, not the people directly.

    How the Electoral College distorts the popular vote

    The electoral vote always distorts the popular vote by exaggerating the winner’s margin of victory. In very close contests, it can also go against the popular vote, as it has done on four occasions – 1876, 1888, 2000 and 2016.

    Two mechanisms are responsible for this.

    First, the populations of small states are over-represented in the Electoral College compared to the larger states because of the guaranteed minimum three electoral votes.

    For example, Alaska, with three electoral votes, has one electoral vote for every 244,463 inhabitants (based on 2020 US census data). In contrast, New York, with 28 electoral votes, has one electoral vote for every 721,473 inhabitants. So, an electoral vote in Alaska is worth almost three time as much as an electoral vote in New York.



    Second, and far more significant, is the “winner-takes-all” arrangement. In every state, except Maine and Nebraska, the winner of the popular vote takes 100% of the electoral votes, no matter how close the contest is.

    Even in Maine and Nebraska, it’s winner-takes-all, except those states award two electoral votes to the statewide winner of the popular vote and one electoral vote to the popular vote winner in each of its congressional districts.

    Few Americans would be conscious of how the winner-takes-all system works, either.

    Put simply, when voters cast a ballot, they are, in effect, voting multiple times – once for each elector in the state supporting the presidential candidate of their choice. They do this by marking just one box alongside their preferred candidate’s name.

    For example, if Harris defeats Trump by 51-49% of the popular vote in Pennsylvania, every one of the 19 electors on Harris’ slate will defeat every one of Trump’s 19 electors by the same margin. The popular vote may have been close, but in the electoral vote, it’s 19-0 for Harris.

    When that is repeated across all 50 states, the Electoral College vote will always exaggerate the margin of victory compared to the popular vote.

    In the 1992 presidential election, for example, Bill Clinton defeated George H.W. Bush by a landslide in the electoral college, 370-168. However, Clinton only edged Bush by 5.5 percentage points in the popular vote (43% to 37.45%). Independent candidate Ross Perot, meanwhile, earned nearly 19% of the popular vote, but because he didn’t carry any states, he got zero electoral votes.

    And when the loser of the popular vote wins the electoral vote, such as Trump’s victory over Hillary Clinton in 2016, it shows the total number of popular votes won by a candidate is less important than where those votes are located.

    To win in the Electoral College, a candidate needs to have their vote distributed economically between the states. In a majoritarian democracy (based on the principle of majority rule), this ought not to be a feature of the electoral system. But the US presidential election process was never designed to operate this way.

    Lastly, the Electoral College also heavily determines the nature of the election campaign. Most states in the US are “safe” wins for one party or the other.

    As such, the efforts of the candidates are concentrated in the handful of states that are competitive – the so-called “battleground” states. The rest of the country tends to be ignored.

    The future of the Electoral College

    That the Electoral College survives into the 21st century is partly due to the adaptability of the Constitution to deal with the earlier challenge in the 1800s over the selection of electors in the states, as well as the immense difficulty of amending the Constitution.

    This is despite the fact a clear majority of Americans support abolishing the Electoral College in favour of a national, direct popular vote for the presidency.

    What happens in this election is anyone’s guess. With the polls showing such narrow margins in the popular vote in the battleground states, the outcome is not only unpredictable, it may even be random. And that’s a terrible comment on the state of American democracy.

    John Hart does not work for, consult, own shares in or receive funding from any company or organisation that would benefit from this article, and has disclosed no relevant affiliations beyond their academic appointment.

    ref. Explainer: a short history of the Electoral College and how it subverts the will of voters – https://theconversation.com/explainer-a-short-history-of-the-electoral-college-and-how-it-subverts-the-will-of-voters-239206

    MIL OSI AnalysisEveningReport.nz

  • MIL-Evening Report: Ancient mud reveals Australia’s burning history over the past 130,000 years – and shows a way through our fiery future

    Source: The Conversation (Au and NZ) – By Michela Mariani, Associate Professor in Physical Geography, University of Nottingham

    Lano Lan / Shutterstock

    Increased land management by Aboriginal people in southeastern Australia around 6,000 years ago cut forest shrub cover in half, according to our new study of fossil pollen trapped in ancient mud.

    Shrubs connect fires from ground cover to the forest canopy, allowing fires to spread and intensify quickly. The reduction in shrub cover, linked to evidence for increasing population size and more widespread landscape use by Aboriginal people, would have dramatically decreased the potential for high-intensity bushfires.

    We also found the shrub layer in modern forests is even greater than it was 130,000–115,000 years ago, when the climate was similar to today’s but there were no people around.

    Our deep-time research shows how important Indigenous cultural practices were for reducing dangerous high-intensity fires. It also suggests a way forward in Autralia’s current fire crisis, which climate change is making worse.

    The trouble with shrubs

    For decades, Australia has tried to manage fires by suppressing them. This strategy may be effective in the short term, but it has led to dire consequences in the long term.

    Over the past 20 years, the forests and woodlands of southeastern Australia have become hotspots for major fires.

    Fire suppression has allowed vegetation, particularly in the shrub layer, to grow without constraint. Shrubby, mid-height vegetation acts as a ladder, enabling fires to spread up from the ground to the forest canopy. This results in more intense and uncontrollable fires.

    Summary timeline of past landscape changes across southeastern Australia. We show changes from pre-human contact (top), through Indigenous population expansion (middle), to the present (post-colonial, bottom).
    Simon Connor, CC BY

    Evidence for denser vegetation comes from tiny, fossilised grains of pollen that are laid down in layers of ancient sediment in wetlands and lake beds. By extracting fossil pollen from mud, scientists can develop a picture of vegetation in the past.

    Our new study used archaeological data and information preserved in ancient mud. We looked at how the vegetation of southeastern Australia changed in response to climate and human management over the past 130,000 years.

    We wanted to see how things changed in key periods: before human arrival in Australia, through periods of Indigenous occupation, and following British colonisation.

    We used sophisticated models to estimate vegetation cover and how it related to human land use at different times.

    Caring for Country

    Indigenous Australians have been the custodians of this continent for millennia. Their journey in Australia started at least 65,000 years ago.

    Direct evidence of cultural burning traces back at least 11,000 years in the Top End, although it may have begun much earlier.

    Indigenous Australian cultural burning practices are complex and varied. However, in many parts of the continent they included regular, controlled burns. These helped to manage vegetation growth and reduce the risk of high-intensity fires.

    Since British colonisation, the landscape of Australia has undergone significant changes, with both more open pastures and more densely vegetated forests. The introduction of European land management practices, including fire suppression, disrupted the fire regimes Indigenous Australians had maintained for thousands of years.

    This suppression-focused approach has led to an accumulation of plant matter, creating a tinderbox ready to ignite.

    A call for change: integrating Indigenous Knowledge

    To address this crisis, a shift in fire management strategies is essential. One promising approach is to integrate Indigenous fire management practices into contemporary fire management plans, working with Traditional Owners to best care for Country.

    This must be done in a way that supports Indigenous livelihoods and fosters connection to Country, not by management agencies simply appropriating Indigenous know-how.

    Indigenous Australians possess hundreds of generations’ worth of experience in managing the country’s fire-prone landscapes. Indigenous-led fire management is already being reinvigorated in northern Australia.

    Our research demonstrates that southeastern forests and woodlands were effectively managed in the past and would also benefit from Indigenous caring-for-Country practices today.

    Reducing dangerous fuels in the shrub layer means less high-intensity fires threatening the bush–urban interface, such as the 2019–20 Black Summer fires.

    Indigenous-led burning at a project site in Tasmania.
    Matthew Newton / RUMMIN Productions

    Higher temperatures and prolonged droughts have created ideal conditions for bushfires to spread. Colonisation has compounded the problems arising from human-driven climate change.

    But there is no fire without fuel. It is the combination of increased biomass and a warming climate that now fuels fires of unprecedented scale and intensity, posing a significant threat to lives, property and ecosystems.

    Australia’s fire crisis is a complex issue that requires a multifaceted solution. By learning from and working with Indigenous practitioners, Australia can develop more effective and sustainable fire management strategies. This collaborative approach offers a path forward to tame the flames and protect the nation’s unique and diverse landscapes.

    Michela Mariani receives funding from the Leverhulme Trust and is affiliated with the Australian Research Council Centre of Excellence for Australian Biodiversity and Heritage (CABAH) and the Centre of Excellence for Indigenous and Environmental Histories and Futures (CIEHF).

    Anna Florin receives funding from the Australian Research Council and is affiliated with the Australian Research Council Centre of Excellence for Australian Biodiversity and Heritage (CABAH).

    Haidee Cadd receives funding from the Australian Research Council and is affiliated with the ARC Centre of Excellence for Australian Biodiversity and Heritage (CABAH).

    Simon Connor receives funding from the Australian Research Council. He is affiliated with CABAH, the Centre of Excellence for Australian Biodiversity and Heritage.

    Matthew Adeleye does not work for, consult, own shares in or receive funding from any company or organisation that would benefit from this article, and has disclosed no relevant affiliations beyond their academic appointment.

    ref. Ancient mud reveals Australia’s burning history over the past 130,000 years – and shows a way through our fiery future – https://theconversation.com/ancient-mud-reveals-australias-burning-history-over-the-past-130-000-years-and-shows-a-way-through-our-fiery-future-239561

    MIL OSI AnalysisEveningReport.nz

  • MIL-OSI USA: ARS Hosts 30th Anniversary Ceremony for USDA George E. Brown, Jr. Salinity Laboratory

    Source: US Agriculture Research Service

    ARS Hosts 30th Anniversary Ceremony for USDA George E. Brown, Jr. Salinity Laboratory

    Contact: Amaani Lyle
    Email: Amaani.Lyle@usda.gov

    RIVERSIDE, Calif., Oct. 31, 2024 — The U.S. Department of Agriculture’s (USDA) Agricultural Research Service (ARS) hosted the 30th anniversary ceremony of the U.S. Department of Agriculture George E. Brown, Jr., Salinity Laboratory on the University of California, Riverside campus today.

    The event celebrated decades of research at the laboratory, where ARS teams have pioneered irrigation practices to optimize crop production, leverage conservation through recycled water, and minimize land degradation caused by the omnipresence of natural and man-made salinization, the overabundance of salt.

    Salinization is the buildup of salts and other trace elements in irrigated soils that reduces the productivity of croplands, constituting a long-standing threat to farming in the semi-arid regions of the American West, parts of the Midwest, and throughout the world.

    The Salinity Laboratory’s innovation and leadership in understanding salt-affected soil-plant-water systems for the conservation and protection of our land and water resources and the maintenance of a viable, permanent irrigated agriculture has garnered acclaim from both USDA leadership and UC Riverside collaborators.

    “This milestone anniversary of the U.S. Department of Agriculture George E. Brown, Jr., Salinity Laboratory marks a storied history of our mission to overcome water quality and water scarcity constraints on agriculture and human health,” said Dr. Simon Liu, ARS Administrator. “Our researchers’ work has yielded palpable solutions to address climate change, drought, and competition for resources — factors that reduce the availability of irrigation water and compound risks posed by salinization.”

    ARS Research Leader Todd Skaggs echoed the importance of appropriate salinity control.

    “To meet increasing demands for food amid limited soil and water resources, the nation and much of the world community will continue to look to the USDA Salinity Laboratory for salinity expertise, water quality research, and applications to solve these problems,” Skaggs said. “Protection of soil resources will become even more vital as water conservation, efficiency and quality assurance continue to increase.”

    According to Skaggs, current research focus areas have earned the lab a vanguard position in addressing and eradicating per- and poly-fluoroalkyl substances (PFAS), chemicals of concern whose presence in recycled water may lead to the introduction of these harmful compounds into the human food chain if such waters are used for agricultural crop irrigation.

    “Even low levels of PFAS may accumulate in soils over time and be absorbed by crop roots,” Skaggs explained. “Our team is therefore developing a low-cost, low-energy, environmentally-friendly treatment system for the removal of PFAS from recycled water, and it’s being optimized for application at the greenhouse scale, potentially allowing for production of PFAS-free crops for human consumption.”

    Skaggs also recounted research inflection points in the 1960s and 1970s when ARS scientists quantified the salinity response of a wide variety of agricultural crops and demonstrated that soil salinity could be managed with significantly less water than used in conventional irrigation practices, setting the stage for substantial water conservation.

    “This research enabled the selection of suitable crops for a given irrigation water quality and the data remain the most comprehensive information available on relative crop salt tolerances,” he said.

    By the 1990s, the Salinity Laboratory developed benchmark methods for mapping and monitoring soil salinity across swaths of U.S. farms and agricultural regions, irrigated lands that produce 30% of the nation’s food and fiber and practically all of the nation’s fruits and vegetables.

    The 30th anniversary ceremony highlighted but a fraction of the lab’s deep roots.

    In 1937, the United States Regional Salinity Laboratory was established at the base of Mount Rubidoux in the city of Riverside before adopting a new moniker, United States Salinity Laboratory, in 1951. 

    In 1986, Congressmen George Brown, Jerry Lewis, and Al McCandless of California introduced legislation to replace the aging U.S. Salinity Laboratory with a modern facility on the UCR campus, which supported the project by leasing to the federal government a 7.5-acre site for 50 years at $1 per year. Groundbreaking occurred in 1992 before the facility officially opened in 1995. 

    In 2000, the laboratory name would change once more to George E. Brown, Jr., Salinity Laboratory in memory of Congressman George E. Brown, Jr.

    The laboratory has the distinction of being the only facility in the country specifically devoted to agricultural and environmental issue analysis through basic research on saline and alkali soils, including related reclamation, irrigation, drainage and soil management.  

    Following the ceremony was an open house highlighting current research including breeding of salt-tolerant alfalfa, precision agriculture, and greenhouse-scale treatment systems for removing PFAS from irrigation water.

    ARS and UC Riverside researchers have maintained collaborative relationships, sharing vital resources to achieve common research objectives through cooperative research programs. 

    Leaders who spoke at the ceremony included:

    • Dr. Simon Liu, Administrator, USDA-ARS 
    • Dr. Tara McHugh, Director, Pacific West Area, USDA-ARS 
    • Dr. Marlen Eve, Deputy Administrator, USDA-ARS National Program Staff, Natural Resources and Sustainable Agricultural Systems
    • Peter Atkinson, Interim Dean, UC Riverside College of Natural and Agricultural Sciences
    • Rien van Genuchten, Recipient of the 2023 Wolf Prize in Agriculture

    The Agricultural Research Service is the U.S. Department of Agriculture’s chief scientific in-house research agency. Daily, ARS focuses on solutions to agricultural problems affecting America. Each dollar invested in U.S. agricultural research results in $20 of economic impact.

    MIL OSI USA News

  • MIL-OSI USA: Neal Announces $850,000 Earmark for Western New England University’s Center for Advanced Manufacturing

    Source: United States House of Representatives – Congressman Richard Neal (D-MA)

    Today, Congressman Richard E. Neal joined Western New England University (WNE) President Dr. Robert E. Johnson, Western New England University Dean of the College of Engineering Dr. S. Hossein Cheraghi, students, faculty, and staff to announce an $850,000 earmark for the establishment of WNE’s Center for Advanced Manufacturing Systems (CAMS).

    The allocation was made possible through Congressionally Direct Spending (CDS) from the U.S. Department of Housing and Urban Development. Congressman Neal included funding for this project in the Fiscal Year 2024 spending bill that was signed into law by President Biden on March 9, 2024. This funding will allow WNE to establish CAMS which will serve as a hub for industry-university collaboration, focusing on training and retraining a workforce in advanced manufacturing techniques.

    “In an ever-changing society fueled by innovation and technological developments, the importance of workforce development cannot be understated. As a result, higher education has come to play a prominent role in training the next generation of workers. That is why I was proud to fight for Western New England University to secure funding that will benefit their students for years to come,” said Congressman Neal. “As one of the top engineering programs in the nation, WNE continues to invest in programs that will lead to immediate job placement upon graduation. This funding will play a critical role in ensuring their students are equipped with the skills needed to meet the demands of our region’s workforce, stimulating economic growth and opportunities.”

    Contributing $2.8 trillion to U.S. GDP in 2023, the manufacturing sector accounts for nearly 12% of the U.S. economy, more than half of which is attributed to advanced manufacturing. Once a manufacturing hub in the northeast, Springfield has witnessed a steady decline in its manufacturing workforce since the early 2000s. This project will help address that decline by revitalizing the sector through partnerships with local companies, public schools, and other higher education institutions to provide workforce training, internships, and research opportunities.

    “We are incredibly grateful to Congressman Neal for his steadfast support and leadership in securing this $850,000 earmark for Western New England University to establish the Center for Advanced Manufacturing Systems (CAMS) within our College of Engineering,” said President Johnson. “With updated facilities and tools, we will enhance our educational experience for students, ensuring that future graduates are equipped to meet the immediate needs of our industry partners. Western New England University remains committed to preparing our students for the future of work and this funding will allow us to stay at the forefront of innovation.”

    Funding for CAMS will allow WNE to launch new academic programs and certifications in advanced manufacturing while continuing to foster partnerships with local manufacturers to drive technology adoption and innovation, including Advance Welding, Nitor Corporation, American Steel and Aluminum Corporation, and Mestek, Inc. These partnerships will support WNE’s goals of establishing CAMS, including:

    • Providing 20 MA companies with access to a state-of-the-art incubator and training laboratory for workforce training on advanced welding and product characterization and prototyping;
    • Incubating two Springfield, MA startup companies a year working with advanced manufacturing techniques and equipment;
    • Educating the current and future workforce in the use of cutting-edge manufacturing equipment, and improving manufacturing processes through four industry-focused workshops a year; and
    • Providing 200 K-12 students an experiential opportunity at the Center around advanced manufacturing processes each year.

    Under guidelines issued by the Senate and House Appropriations Committees, members of Congress requested CDS funding for projects in their state for Fiscal Year 2024. CDS requests were restricted to a limited number of federal funding streams, and only state and local governments, and eligible non-profit entities, were permitted to receive CDS funding.

    This project is one of thirteen CDS projects submitted by Congressman Neal, totaling nearly $15 million in investments throughout the First Congressional District of Massachusetts.

    ###

    MIL OSI USA News

  • MIL-OSI Australia: On cloud eight: UniSA secures grants to support struggling kids

    Source: University of South Australia

    01 November 2024

    Whether it’s the development of a board game to tackle teenage mental health, or deciphering treatment-resistant brain tumours, the University of South Australia is stepping up to support children struggling with mental and physical health concerns.

    Eight UniSA research projects have been awarded almost $740,000 in research funding from the Channel 7 Children’s Research Foundation (CRF) to help improve children’s health and wellbeing.

    The successful projects comprise:

    • MindScape: Co-creation of an evidence-based board game to address adolescent mental health. Lead investigator: Dr Jacinta Brinsley.
    • A Phase 1 trial of group improvisational theatre games that include Acceptance and Commitment Therapy treatment for children and teens who stutter. Lead investigator: Dr Michelle Swift.
    • Minimising treatment-induced disease in children with the nerve cell cancer neuroblastoma. Lead investigator: Professor Greg Goodall.
    • Turning off the danger signal: reducing ventilator-induced lung injury in preterm babies. Lead investigator: Dr Jack Darby.
    • Launching Early Childhood Numeracy: Fostering mathematical self-efficacy of preschool children and their parents for lifelong learning. Lead investigator: Dr Chelsea Cutting.
    • Developing lab-grown paediatric brain tumours that mimic the response to treatment of real tumours to learn how they adapt and resist radiotherapy. Lead investigator: Professor Benjamin Thierry.
    • Improving the understanding of fetal DNA found in maternal blood during pregnancy towards designing more comprehensive non-invasive prenatal genetic tests. Lead investigator: Dr Marnie Winter
    • Re-examining the scope of statutory child protection to improve responses to children. Lead investigator: Professor Leah Bromfield.

    UniSA Deputy Vice Chancellor Research and Enterprise and Standing Acting Vice Chancellor, Distinguished Professor Marnie Hughes-Warrington AO, says the support provided by the Channel 7 CRF is invaluable.

    “The Channel 7 Children’s Research Foundation is a longstanding supporter of quality research into the causes, prevention, diagnosis and treatment of complex health and welfare issues facing children today,” Prof Hughes-Warrington says.

    “All children deserve to thrive. These grants will enable our researchers to investigate, trial and uncover new solutions so that young children can live fuller and healthier lives.

    “We are very proud of this achievement, our best ever result with the Foundation, and look forward to continuing invaluable work for children.”

    …………………………………………………………………………………………………………………………

    Media contacts:
    Annabel Mansfield M: +61 479 182 489 E: Annabel.Mansfield@unisa.edu.au
    Maddie Rawlings E: Maddie.Rawlings@unisa.edu.au

    Other articles you may be interested in

    MIL OSI News

  • MIL-OSI Economics: Per Jacobsson Lecture 2024 — Ngozi Okonjo-Iweala: “Delivering on new global challenges: How can we keep multilateral coherence whilst re-imagining the multilateral trading system?”

    Source: WTO

    Headline: Per Jacobsson Lecture 2024 — Ngozi Okonjo-Iweala: “Delivering on new global challenges: How can we keep multilateral coherence whilst re-imagining the multilateral trading system?”

    Excellencies, Dear Raghu, Minouche, Maury, ladies and gentlemen, friends,
    Thank you. What an honor to follow in the footsteps of previous Per Jacobsson lecturers – all the more so in this 80th anniversary year of the Bretton Woods Conference.
    We are living in troubled times – something Per Jacobsson knew well. So far as trade is concerned, the times are not only troubled, they are tense. Trade is sometimes blamed and scapegoated for poor outcomes that really derive from macroeconomic, technology, or social policy, for which trade is not responsible.
    Trade policies and tools are being deployed not just to solve trade-related problems, but also to try to address security and geopolitical concerns.
    As unilateral measures or threats thereof become increasingly widespread, trade policy has been getting more restrictive. In recent months, the US, the EU, Turkey, and Canada have introduced new tariffs and countervailing duties on Chinese electric vehicles and other products, including steel. China has countered with WTO disputes and measures against EU products such as dairy, pork, and brandy. 
    These are among the over 130 new trade-restricting measures recorded by the WTO Secretariat since the start of this year. This number represents an 8% increase to the stockpile of over 1600 restrictive measures introduced between 2009 and 2023, which as of last year were already affecting over 10% of world goods trade. In addition, WTO members initiated 210 trade remedy investigations in the first half of 2024 – nearly as many as in all of 2023. While not all will culminate in the imposition of duties, investigations have a well-documented chilling effect on trade. And I haven’t even mentioned subsidies yet. 
    Frictions are manifesting as trade disputes. Six of the eight WTO disputes initiated this year deal with green technologies, particularly electric vehicles.
    I hope we are not on a path that leads back to the sort of economic disorder that came before Bretton Woods – disorder that was followed by political extremism and war.
    It was precisely to avoid a repeat of such circumstances that the multilateral economic institutions were created. My concern today is that we have forgotten this lesson – that we have forgotten the good these institutions have done.
    Walking away from the legacy of Bretton Woods, including the trading system, would diminish the world’s ability – collectively and at the national level – to respond to problems affecting people’s lives and opportunities.
    I will argue that there is a better path forward: re-imagining the global trading system and the rest of the multilateral economic architecture to help us meet the technological, environmental, social and geopolitical challenges of our time. To succeed, its various components must work in concert – an idea we have come to call ‘coherence’.
    In the 1940s, the overall thrust of coherence was that trade, reconstruction financing, and monetary policymaking need to be in harmony with each other, and anchored in institutions and rules across countries, to promote growth, prosperity, and peace.
    Today, delivering lasting improvements to people’s lives and livelihoods requires us to solve problems of the global commons.
    The notion of coherence across different policy areas would have made sense to Per Jacobsson. His convictions about sound money, and its importance for durable growth and recovery, were shaped by his own experiences. As a young man he saw the collapse of global economic integration amid the First World War. From his position at the League of Nations in the 1920s, he witnessed the failed attempts by leading economies to establish effective international coordination on global finance and trade – a memory that echoes uncomfortably today.
    We know what happened when the downturn came at the end of the decade. Vicious circles emerged: of falling output, deflation, banking and financial crises, trade protectionism and retaliation, and exchange rate chaos. Countries retreated into increasingly isolated economic blocs.
    The experience of those years was seared into the consciousness of the officials who gathered in Bretton Woods in July 1944. US Treasury Secretary Henry Morgenthau opened the conference by looking back at what he called “the great economic tragedy of our time.” I quote “We saw currency disorders develop and spread from land to land, destroying the basis for international trade and international investment and even international faith. In their wake, we saw unemployment and wretchedness — idle tools, wasted wealth. We saw their victims fall prey, in places, to demagogues and dictators. We saw bewilderment and bitterness become the breeders of fascism and, finally, of war.”
    What Bretton Woods delivered
    The genius of Bretton Woods was that it turned the vicious circles of the 1930s into virtuous ones, by recognizing that macro-financial stability, reconstruction and development, and trade went hand-in-hand.
    Instead of beggar-thy-neighbor policies, countries would treat trade, monetary issues, and even domestic macro-economic policies as matters of common interest.
    Instead of excessively rigid or chaotically fluctuating currencies, there would be orderly, rules-based management of exchange rates and balance of payments problems.
    Instead of underinvestment, there would be long-term financing for reconstruction and expanding productive capacity.
    Instead of quantitative restrictions, prohibitive tariffs, and bilateral clearing, there would be a coordinated lowering of trade barriers, and freedom to undertake international payments and current account transactions.
    The idea of coherence across policy fields, with trade as a unifying theme, was baked into the system from day one. Promoting the “balanced growth of international trade” is written into the founding mandates of both the IMF and the World Bank – not as an end in itself, but as a means to higher employment, productivity, and incomes.
    The trade leg of the stool, alongside the Bank and the IMF, was supposed to be the International Trade Organization, but it ran aground in the US Congress. A parallel negotiating process in 1947 produced the General Agreement on Tariffs and Trade, which was nominally temporary and did not require Congressional ratification. Successive rounds of GATT negotiations substantially reduced barriers to trade. The growing number of “contracting parties” used the GATT to resolve and avoid trade disputes. By the 1960s, global trade was growing faster than output.
    The decades that followed Bretton Woods and the Marshall Plan delivered a breathtaking recovery from the devastation of the Second World War.
    Strong growth in the 1950s and 1960s saw per capita incomes in Western Europe and Japan begin to converge with those in the United States.
    Major European currencies achieved full convertibility in 1958, when Per Jacobsson was leading the IMF.
    These gains, however, were largely confined to industrialized countries.
    Most newly independent developing countries continued to lose ground in relative terms, as they struggled with declining terms of trade for their commodities.
    But a handful of poor economies in East Asia started trying to use increasingly open external markets to pursue export-led development.
    Discordance and reinvention: the 1970s and 1980s
    Coherence gave way to discordance in the 1970s, with the oil shocks, stagflation, the advent of floating exchange rates, and a wave of emerging market debt crises.
    By the mid-1980s, the success of the so-called Asian tigers had become a compelling example, inspiring many developing country governments to pivot from inward-oriented to export-oriented development strategies.
    At the international level, growing frustration with ad hoc protectionism and “à la carte” approaches to GATT strictures created demand for more rules-based trade cooperation.
    The Uruguay Round negotiations from 1986 to 1994 broadened the reach of multilateral trade rules to cover services and intellectual property, filled longstanding gaps with respect to agriculture and textiles, and unwound much of the protectionism that had emerged in the preceding years.
    The nominally provisional GATT was transformed into the World Trade Organization, with a binding dispute resolution mechanism that enhanced the predictability offered by its expanded rulebook.
    The preamble to the Marrakesh Agreement establishing the WTO opened up new vistas for the organization, defining its purpose as using trade not just to raise living standards and create jobs but to advance sustainable development – thus introducing environmental concerns that were absent in the 1940s.
    1990 to 2020: A “golden period of economic development”, but clouds on the horizon
    The Uruguay Round and the end of the Cold War would mark a second era of coherence and virtuous circles across the trading system, the World Bank, and the IMF. And this time, the benefits were spread much more widely across countries and people.
    The WTO became an anchor for outward-oriented economic reforms in many emerging markets and developing economies.
    Increasingly open and predictable trade became a stronger driver of development, productivity, specialization and scale.
    Better macro-financial policies bolstered growth – and trade performance – in many emerging markets and developing countries. So did improved human capital and physical infrastructure.
    Trade and modern supply chains became powerful sources of disinflationary pressures.
    Market-oriented reforms in China, Eastern Europe, India and other developing economies brought them into the increasingly global division of labor. Trade boomed, incomes rose, and poverty plummeted.
    Between 1995 and 2022, as low- and middle-income economies nearly doubled their share in global exports from 16 to 32%, the share of their populations subsisting on less than US$2.15 per day fell from 40% to under 11%. Over 1.5 billion people were lifted out of extreme poverty.
    Since 1995, per capita incomes in low- and middle-income countries have nearly tripled, and global per capita income increased by approximately 65 percent.
    For the first time since the industrial revolution two centuries earlier, per capita incomes in rich and poor countries began to converge.
    Gains for poor countries did not come at the expense of rich ones. Examining the United States since 1950, researchers at the Peterson Institute for International Economics (PIIE) have shown that international trade boosted the economy by the equivalent of $2.6 trillion in 2022, or about 10% of GDP. The gains from trade would be even larger for small, open advanced economies.
    In a Foreign Affairs piece this year, Dev Patel, Justin Sandefur, and Arvind Subramanian called the years between 1990 and the start of COVID-19 pandemic in 2020, I quote, “history’s most golden period of economic development”.   They argue that the rapid increase in trading opportunities was “perhaps the most important enabler” of convergence.
    Research from our new World Trade Report backs them up: the pace of income convergence of low- and middle-income economies is strikingly correlated with their participation in global trade, as measured by a size-adjusted ratio of trade to GDP. Our simulations suggest falling trade costs account for as much as one-third of the convergence.
    To be clear, the period was not golden for everyone. Developing countries with lower trade participation or greater commodity-dependence – mostly in Africa, Latin America and the Caribbean, and the Middle East – lagged on convergence. And in some rich countries, many people felt left behind, and their frustration started to fuel a political backlash against trade.
    Multilateral rule-making on trade began to falter, with the failure of the Doha Round of WTO negotiations.
    Nevertheless, in 2008 and 2009, when the world economy faced its worst financial crisis since the 1930s, the system worked.
    International markets stayed broadly open. The rules and norms of the multilateral trading system helped governments contain protectionist pressures.
    Alongside fiscal and monetary support, trade was a powerful shock absorber. Crisis-hit countries could rely on predictable market access elsewhere to absorb their excess supply, preventing growth and development from getting derailed.
    The WTO, the World Bank, and the IMF also worked together productively on the macro-micro policy nexus.
    For instance, when trade finance dried up during the credit crunch, despite being extremely low-risk, the three institutions joined hands to encourage G20 members and international financial institutions to step in with a $250 billion support package.
    Since the financial crisis, the multilateral trading system, with the WTO at its core, has continued to deliver economic benefits, despite rising geopolitical tensions and tariffs between the US and China, the disabling of the Appellate Body, and the failure to reach agreements in long-running negotiations such as those on agriculture. Global trade kept reaching new highs through the 2010s, and over 75% of global goods trade continued – and continues today – to operate on core WTO tariff terms.
    When COVID-19 hit in 2020, the norms and rules of the multilateral trading system mostly did their job again. Trust in trade was damaged by initial missteps, as governments enacted export restrictions on medical supplies and vaccines. But governments generally refrained from widespread protectionism, allowing food and other essentials to flow across borders to where they were needed. Goods trade rebounded strongly from the lockdowns and was soon setting new records. Cross-border supply chains churned out products needed to fight the pandemic, from face masks to vaccines. Trade in digitally-delivered services boomed, propelled by the same technologies that allowed so many of us to work from home.
    Goods and especially services trade are now well above pre-COVID levels.  Last year, global trade was worth a near-record $30.5 trillion, in a $105-trillion world economy.
    Re-imagining the Multilateral Trading System with coherence
    As we saw at the outset, however, these successes did not forestall the challenges we now face in global trade. While trade has been largely resilient, signs of fragmentation are now visible.
    So it’s not difficult to imagine a return of vicious circles – trade restrictions, efficiency losses, slower growth, higher prices, costs imposed by extreme weather and food insecurity, and public frustration and anger.
    Allowing the vicious circles to take hold and the world to fragment into isolated trading blocs would be costly. The WTO has estimated longer term global GDP losses in the order of 5% were the world to fragment into two like-minded trading blocs. IMF estimates are in the order 7%. We cannot afford this!
    And that is why we need to re-imagine the multilateral trading system to solve modern challenges and address modern vulnerabilities.
    This means re-imagining coherence as well. Trade alone was insufficient in 1944, and trade alone is insufficient to build the more secure, sustainable, and inclusive world we want today.  The way forward for trade will increasingly be about “WTO and” – trade in tandem with other issues, and policies that support the original vision of coherence and do not misuse trade tools, for coercion, as a weapon, or to undermine competition.
    Our unfinished business from 1944 was elegantly illustrated by a recent blog post from IMF chief economist Pierre-Olivier Gourinchas and his team.
    They showed that China’s growing and contentious trade surplus, and the US’s widening trade deficit, are the result of domestic macro-economic forces, rather than the product of trade and industrial policies.
    “Homegrown surpluses and deficits call for homegrown solutions,” they argued, suggesting demand-boosting measures in China and fiscal consolidation in the US.
    As for concerns over industrial policy, they said the right response was to strengthen WTO rules, not to restrict trade.
    They cited the WTO’s recent China Trade Policy Review which showed new data of billions of dollars in subsidies going to manufacturing. Urging China to be more transparent about its subsidies.
    The blog shows the coherence mandate in action but it also illustrates how even today, the global trading system is paying a price for shortcomings of macro-economic policy.
    As Sylvia Ostry, one of my predecessors at this podium, said in 1987, “Trade policy is no substitute for macro policy.”
    Let’s now turn to the new trade agenda, and look at three areas where future prospects for people and the planet require trade to be re-imagined, and complemented by other policy levers pulling in the same direction.
    First, the environmental agenda, above all climate change and getting to net zero by mid-century.
    Trade is indispensable to deploy low-carbon technologies globally. Trade lets countries share the burden of developing new green tech. Scale economies and competitive pressures associated with trade help drive down unit costs, making it possible for renewables to undercut fossil fuel energy.
    Trade also allows us to leverage ‘green comparative advantage’, a concept that our chief economist, Ralph Ossa, has done much to advance. The idea is straightforward: just as individuals and countries can reap economic gains by specializing in what they are relatively good at, the world can reap environmental gains if countries specialize in what they are relatively green at.
    If countries with abundant clean energy can produce more energy-intensive goods and services, while importing energy-light products from places where clean energy is scarce, and vice versa, global emissions fall much more than they would have absent that trade. And in fact research from the University of Zurich  suggests that as much as one-third of global emissions reductions could come from this kind of specialization linked to green comparative advantage.
    As Ricardo Hausmann at Harvard has observed, fossil fuels are cheap to transport, but wind and solar energy are not. This makes parts of Africa, Central Asia, and Latin America with high green energy potential attractive destinations for investment in energy-intensive industries, including the production of green hydrogen.
    Global cooperation on internalizing carbon costs would incentivize greener sourcing everywhere. Nevertheless, we are already seeing moves in the right direction as in Kenya, which has attracted a billion-dollar investment to build a geothermal-powered low-carbon data center.
    Parenthetically, a similar dynamic exists for water, provided it is valued correctly. A recent report of the Global Commission on the Economics of Water, which I co-chair, shows that with trade one can also promote the notion of a hydrological comparative advantage. Trade can help mitigate water scarcity by allowing countries with abundant hydrological resources to specialize in producing water-intensive products for export to water-scarce nations.  Such virtual water trade offers agricultural export opportunities, for example, to those regions including countries in Africa with under-utilized ground water resources and land.
    But just as environmental policy coordination could accelerate climate action, policy fragmentation could weaken it.  There is a genuine risk that trade frictions associated with carbon pricing, green subsidies, and other climate policies will escalate into trade restrictions and retaliation, harming emissions reduction as well as trade.
    We should seek to pre-empt such frictions and disputes by establishing shared frameworks for trade and climate policy. The goal would be to maximize emissions reduction and green innovation, while minimizing negative spillovers, trade tensions, and wasted public resources on subsidy races that most countries may not even afford to participate in.
    To this end, the WTO Secretariat is coordinating a carbon pricing task force comprised of the IMF, World Bank, OECD, UNCTAD, and UNFCCC, where we are working to develop shared carbon metrics and ultimately a global carbon pricing framework against which we can benchmark national policies to aid interoperability of approaches. We have also joined hands with the IMF, the OECD, and the World Bank to explore approaches to enhance greater transparency with respect to subsidies. And we are working with the steel industry to help them promote interoperability in decarbonization standards, reducing transaction costs and facilitating trade and investment in green steel.
    Reforming the over $1.2 trillion in direct global annual fossil fuel subsidies, the $630 billion in trade-distorting agricultural support, and the $22 billion in harmful fisheries subsidies (which the WTO Fisheries Subsidies Agreement is delivering) should be a no-brainer. Some of the resources freed up could be repurposed to support green innovation and a just transition for poor countries.
    The second set of opportunities for the Multilateral Trading System deals with diversifying and decentralizing supply chains – and doing so in a manner that brings in countries and communities that remain on the margins of the global division of labor.
    More diversified global production networks would enhance supply security in an increasingly shock-prone world, while extending the benefits of trade to places and people that have not shared adequately in them. Greater diversification would also help lower the geopolitical temperature around supply chain relationships, by making them harder for any single country to weaponize.
    As the pandemic and the war in Ukraine made abundantly clear, overconcentration makes supply chains vulnerable in a crisis.
    The advent of COVID-19, concentrated minds on the fact that 80% of world vaccine exports came from only ten countries. This meant export restrictions in a few of them severely disrupted global access to vaccines – especially to Africa, which relied on imports for 99% of its jabs.
    Decentralizing value chains and building up pharmaceutical production capacity in Africa and other developing country regions for instance would make the global supply base more resilient in the event of future pandemics, whilst more closely integrating these regions in to world trade, and making them part of a more prosperous and healthy world.
    Critical minerals is another sector where there are major opportunities to mitigate concerns about overconcentration in mining and especially processing, while stimulating growth in developing countries. 
    Exports of minerals critical for the low-carbon transition, like lithium, cobalt, nickel, and rare earths, have grown rapidly to reach USD 320 billion in value in 2022, and are set to increase much more in the years ahead. Africa, for example, represents 40% of estimated global reserves of cobalt, manganese, and platinum; and 12% of world exports of critical minerals, but only 3.8% of exports of processed minerals.
    By investing in processing these minerals within the regions including in Central Asia and Latin America where they are found, we can promote value addition and job creation while removing supply bottlenecks that currently threaten to hold back the low-carbon transition.
    Furthermore, to the extent that this process is powered by green hydrogen and other kinds of clean energy, it would harness the green comparative advantage I mentioned earlier and thereby help the developing regions increase their share in world trade.
    It would be green growth and green trade – the ‘re-globalization’ we want.
    Finally, there are areas where cross-border commerce is flourishing, but where new rules are necessary to foster predictability and lower barriers to entry for smaller businesses and developing economies.
    The fastest growing segment of international trade is in services delivered across borders via computer networks. Trade in digitally-delivered services – everything from streaming video to remote consulting – has quadrupled since 2005, reaching $4.25 trillion in value last year. These services have become an increasingly important driver of growth and job creation.
    The commercialization of artificial intelligence promises to further accelerate digital trade. A forthcoming WTO report describes how AI could reduce trade and transaction costs, improve supply chain logistics, and shift countries’ comparative advantages.
    I always say the future of trade is digital, but the future of protectionism could be as well. Imports of digital services could become as contentious as manufactured imports have, or more so – inviting digital barriers that are even simpler to put in place than their counterparts for trade in physical goods.
    Putting in place some basic rules for digital trade would reduce the risks of such reversals. The 90-odd members participating in plurilateral e-commerce negotiations at the WTO are now looking to conclude a first phase agreement on a series of practical measures to facilitate digital trade, from common rules for e-signatures and payments, to paperless trading, and consumer protection. Tougher issues like cross-border data flows – a critical element in AI – will be dealt with in a second phase of negotiations.
    Delivering on this agenda for the future will involve strengthening all of the WTO’s functions: monitoring and transparency, negotiations, and dispute settlement.
    With respect to our dispute settlement system, we are working to reform it. The reform process has wide buy-in, and talks are advancing, including on issues like appeal review and accessibility to ensure that developing countries can use the system. There are delicate issues here around how national security exceptions will be handled – it is going to take work!
    We will need to negotiate and implement new rules in important areas like the environment. Some members are showing the way: New Zealand, Costa Rica, Switzerland, and Iceland recently agreed to liberalize trade in a list of hundreds of environmental goods, and they are trying to get others to join.
    We are working on getting an Agreement on Investment Facilitation for Development, negotiated by three-quarters of our membership, into the WTO rulebook. This agreement will help developing economies attract FDI by simplifying investment-related procedures and sweeping away red tape.
    We will also need to review existing rules to make them fit for purpose. Instead of members doing an end run around our Agreement on Subsidies and Countervailing Measures to introduce industrial policies, it would be better to update that agreement. It actually dates back to 1994 – seven years before China joined the WTO,  [a time when climate concerns were barely on the radar screen, and the conventional wisdom was that state-owned enterprises were a fading relic of a bygone era]. Members could decide to create space for subsidizing the green transition. Shared ground rules would help minimize negative spillovers and related trade tensions, while maximizing efficiency in the use of public resources. 
    Excellencies, ladies, and gentlemen. Let me now conclude.
    As I said at the start, these are tense times for trade. There are political dynamics outside our control. But we can treat the challenges we face as opportunities to re-imagine the global trading system.
    We can build global resilience whilst making the system more supportive of inclusive growth and environmental sustainability.
    We can make existing trade rules more fit for purpose rather than go around or against them and we can make new rules fit for the time.
    We can help developing countries left behind by the recent wave of global economic integration.
    We can have interdependence without overdependence.
    While nothing is ever easy at the WTO, we are moving in the right direction. We will manage what we can manage. Control what we can control. But we will need your help.
    Over the past eight decades, the multilateral economic architecture, including the trading system, has delivered a great deal for the world. We have reinvented it before. We can do so again, for people and planet.
    Nelson Mandela once wrote that “after climbing a great hill, one only finds that there are many more hills to climb.” I ask you, let’s climb these hills together.
    Thank you.

    Share

    MIL OSI Economics

  • MIL-Evening Report: 5 things you can do to end the biodiversity crisis as the world talks about it at COP16

    Source: The Conversation (Au and NZ) – By Jim Radford, Associate Professor, Ecology and Environment, La Trobe University

    The world is charging towards tipping points for species extinctions, ecosystem collapse and loss of genetic diversity. Crossing these tipping points will be devastating for nature and human existence alike.

    Avoiding this catastrophe of humanity’s making is the purpose of the 16th Conference of the Parties to the United Nations Convention on Biological Diversity (COP16) in Cali, Colombia. COP16 has been reviewing progress on implementing the Global Biodiversity Framework adopted at COP15 in Montreal, Canada, in 2022. Progress has been incremental at best.

    These pledges, plans and goals, while necessary and commendable, are also far removed and often intangible for everyday citizens. Collective global action is inherently political. It moves at glacial pace when urgent action is needed.

    The issues can seem so colossal and complex that individuals often feel powerless. This may mean they do nothing or, worse, add to the problem. But, in fact, there are five steps individuals can take to help end the biodiversity crisis.

    So why isn’t government action enough?

    COP16 wraps up on November 1, but has so far failed to live up to expectations. The COP16 chair claims it has put biodiversity “on an equal footing” with climate. However, solid commitments have yet to emerge.

    For example, before COP16, governments had pledged only US$250 million (A380 million) of the estimated $200 billion per year required by 2030 for the Global Biodiversity Framework Fund. Pledges of another $163 million this week take the total number of contributors to a mere 12.

    Only 15% of countries (including Australia) met the deadline to submit their plans to meet the goals set at COP15. These include protecting at least 30% of the world’s land and water and restoring 30% of degraded ecosystems by 2030.

    And plans do not guarantee action. Indeed, the world has never achieved a single global nature target set by such initiatives.

    Our everyday decisions can’t be divorced from nature

    “Natural capital” is a buzzword in global initiatives, government policies, marketing slogans and sustainability frameworks worldwide. Natural capital refers to all living and non-living natural resources that provide products and services of value to society. In essence, it’s what we commonly call “nature”.

    Understanding and managing natural capital is crucial for conserving biodiversity, addressing climate change and ensuring future generations’ wellbeing by not exceeding our planetary boundaries. It’s why we’ve recently created the Natural Capital Primer. It’s a website that explains how our everyday lives, businesses and economies depend on nature.

    By understanding our connection to nature, we can all reduce our impact on nature. Here are five ways you can make a difference, starting today.

    The Natural Capital Primer explains the concept, aiming to shift attitudes toward nature and promote global conservation.

    1. Cut consumption when you can

    Do you really need to update your mobile phone, your summer wardrobe or your flat-screen TV? What we buy reverberates around the globe.

    Our demand for new products affects resource extraction (leading to habitat loss), carbon emissions (propelling climate change) and pollution (degrading habitat). These impacts are often far from where we make our purchases. From the lithium in our phones to the plastics in our clothes and the metals in our vehicles, our consumption drives demand, which almost inevitably harms biodiversity.

    If you do need to replace something, consider buying second-hand or products made from recycled materials.

    2. Watch what you eat

    Agriculture is the single greatest driver of changes in land use and biodiversity loss. We all need to eat, of course, but where possible buy local and sustainably produced foods.

    Reducing processed foods in your shopping trolley is a good start. Cutting your intake of over-fished, wild-caught seafood, red meat and palm oil-based products will also help. This issue is not straightforward because these products are available as a confusing mix of unsustainable and sustainable options.

    A further complication, made worse by the rise of greenwashing, is that it can be hard to work out exactly what is in certain foods or where they came from. Sustainability certification and apps (GoodFish Australia, for example) can help consumers make better choices.

    3. Choose renewable energy

    The climate and biodiversity crises are inseparable. Neither can be resolved in isolation. For example, nature-based solutions, such as protecting forests as carbon sinks, will help with both the climate crisis and biodiversity.

    With greenhouse gas emissions driving climate change, which threatens many species, a whole range of our choices determine the impacts of our energy use. From your mode of transport to powering your home, choose renewable energy sources.

    Tech giants such as Google and Amazon are turning to nuclear energy to power their generative AI and cloud storage in an effort to reduce their climate impact. However, 100% renewable energy is realistic if consumers demand it from their power companies and governments.

    4. Get your hands dirty

    You can take direct action to protect and increase biodiversity. Volunteer or donate to environmental projects in your neighbourhood. Not only will this make you feel good, but revegetation and habitat restoration do improve local biodiversity.

    Many grass-roots, community-driven projects are making a difference on the ground. They range from urban restoration work, such as the Merri Creek restoration in Melbourne, to forest stewardship projects, such as Tarwin River Forest in Gippsland, Victoria. Get local and get involved!

    5. Adjust expectations and accept responsibility

    People in wealthy countries (such as Australia) have both the biggest environmental footprints and the most capacity to adapt. They must lead change.

    The process starts with increasing awareness of the issues and taking responsibility for change. That includes adjusting our expectations about how and where we live.

    Small changes are magnified when repeated by millions of people. We should never doubt the power of cumulative impact. After all, it’s what got us into this mess in the first place.

    So while governments and corporations haggle, posture and delay over global targets and policies, we can all start right now to make a difference through smarter decisions and sustainable choices.

    Jim Radford receives funding from Australian Department of Climate Change, Energy, Environment and Water, the National Environmental Science Program Resilient Landscapes Hub, Transport for NSW, SmartSat CRC, Macdoch Foundation and Australian Wool Innovation. He is a member of Standards Australia Biodiversity Committee and North Central CMA Science Advisory Panel.

    ref. 5 things you can do to end the biodiversity crisis as the world talks about it at COP16 – https://theconversation.com/5-things-you-can-do-to-end-the-biodiversity-crisis-as-the-world-talks-about-it-at-cop16-242205

    MIL OSI AnalysisEveningReport.nz

  • MIL-OSI USA: Bergman Leads on Call for Election Integrity in Michigan

    Source: United States House of Representatives – Congressman Jack Bergman (MI-1)

    Today, Representative Jack Bergman led Michigan Republican Members of Congress in sending a letter to Michigan Secretary of State Jocelyn Benson demanding answers on Michigan’s elections and urging her to protect election integrity.

    Following reports of a Chinese national University of Michigan student – who is not a U.S. citizen – casting a ballot for the upcoming Presidential election, Rep. Jack Bergman, joined by Reps. John Moolenaar, Bill Huizenga, Tim Walberg, John James, and Lisa McClain, wrote a letter to Secretary Jocelyn Benson outlining their concerns and demanding answers on this issue.

    You can read the letter here or below:

    Secretary Benson:

    As you know, the State of Michigan will play a pivotal role in next week’s election, potentially deciding who will become the 47th President of the United States and which party will control the United States Congress.

    On October 22, 2024, you made the following statement: “Let me be clear: We have secure elections in Michigan.” Furthermore, you testified before the Committee on House Administration on September 11, 2024, stating there is “no evidence non-citizens are voting.” And yet, according to a recent report by The Detroit News, a University of Michigan student who is a Chinese national – and not a U.S. citizen – allegedly cast a ballot for the upcoming presidential election. The report goes on to note that his ballot will be counted.

    Only U.S. citizens are legally permitted to vote in federal elections. You acknowledged this fact during a recent interview on CNN, noting that “Only US citizens can vote in our elections, and secure processes are in place in every state to ensure and reinforce that.” Clearly, this acknowledgement is at odds with the aforementioned report.

    Having instances of non-citizens voting, however isolated they may be, puts doubt into the minds of the millions of Michiganders who entrust you with ensuring our elections are secure, free, and fair, and that the results are valid.

    With this in mind, we would like you to answer the following questions:

    1. What steps is your office taking to verify the citizenship and residency of individuals registering to vote in Michigan to ensure only eligible voters participate in the upcoming election?

    2. Will you commit to prosecuting ineligible voters, like the Chinese University of Michigan student, to the fullest extent of the law?

    3. How is the Michigan Secretary of State’s office addressing potential risks related to ineligible individuals voting, and what security measures are in place to protect the integrity of the voter rolls?

    4. What avenues of recourse are available to ensure improper ballots are not counted? And will you pursue them?

    Please respond no later Friday, November 1, 2025, at 5 p.m.

    MIL OSI USA News