Connect with us


Are Chatbots the technical answer to flattening the Coronavirus curve?




The novel coronavirus is keeping the world under a tight grip. Governments from across the world are finding newer measures to flatten the curve, schools are getting closed, companies are finding newer ways to keep their profits afloat.

While on one side, the economies are taking efforts to lower the impact on coronavirus on industries, new-generation technologies, on the other hand, are taking efforts to make the process of curbing the spread slower.

In this article, we are going to look into a technology that is making its fair share of efforts to flatten the curve — Chatbots.

Use Cases of Chatbots Bringing the COVID-19 Spread Down

A. Dissemination of Information

One of the biggest challenges of every pandemic is the abundance of information that comes sweeping in from all directions. It is the same story all the time. Accurate information comes much later than fake information. Chatbots — taking information from authorized public health agencies — can help bring this issue down. They can offer people concise information which has been shared by the right authorities.

B. Monitoring of symptoms

Considering the nature of the coronavirus, the healthcare frontlines have to be protected a lot more than people self-quarantining. In order to make it successful, it would be important to give people a platform to monitor themselves. This is where chatbots come in handy. By making people answer simple questions, bots will be able to help the users monitor themselves.

C. Mental Health Support

Even though the national and global health bodies are highlighting the need of mental health in a pandemic, the frontlines are, however, not extremely equipped in offering psychological support. The chatbots, can be a great way to eliminate the psychological issues of isolation by becoming 24*7 listeners to the users.

1. Knowledge graphs and Chatbots — An analytical approach.

2. Blender Vs Rasa open source chatbots

3. Sentiment Analysis Voice Bot

4. Chatbot, Natural Language Processing (NLP), Search Services and how to mash them up for a better user experience

Chatbots are a great digital medium for lowering the load on the healthcare domain. By enabling people to self-serve and monitor themselves and backing them with accurate and relevant information, the hospitals are able to provide help and care to the ones who need it.

But the use cases of chatbots in the fight against COVID-19 is not just limited to the health and care domain. They are also being used in multiple scenarios by other industry verticals.

Volunteering- More people want to get involved in the coronavirus-fighting community and help. Chatbots are being employed by NGOs across the globe for making the “How to get involved” information accessible in real-time, with less human interaction.

Retail: The retail houses are changing their hours of operation and limiting shipping options because of the closures. Keeping people updated about their product delivery status is a responsibility that these retail houses are outsourcing to chatbots.

Restaurants: A number of restaurants around the world are limiting the number of employed staff. They have now started incorporating chatbots in their systems for the purpose of taking in orders from their customers.

Chatbots That Are the Textbook Examples of How the Technology Can be Used to Limit the Coronavirus Spread

1. Rapid Response Virtual Agent Program

Announced by Google, the bot enables Google Cloud customers to respond to questions coming in from their customers more quickly. The bot is mainly designed for the organizations who wish to offer COVID-19 related information to their customers — these include government bodies, financial, retail, or travel industries.

2. MyGov Corona Helpdesk

It is the official chatbot of the Government of India, launched to answer questions around COVID-19. It has been made available on WhatsApp. The bot provides huge information about the pandemic in an easy to understand, conversational manner, which includes symptoms, health & travel advisories, and transmission details.

3. Microsoft Azure

The tech giant is offering companies and the healthcare organizations focusing on COVID-19 an infrastructure to develop enterprise-level chatbots which can interact, listen, and understand. The technology has also been devised in a way that it can automatically answer queries about COVID-19 and evaluate the patients before inviting them for in-person health checks.

4. Cobot-19

The bot has been developed by Verloop and Portea Medical to help people in accessing information surrounding COVID-19. All the information that the bot is offering has been gathered from the Centers for Disease Control and Prevention, WHO, and the John Hopkins University.

Cobot-19 too functions through WhatsApp where people can choose from several available options on the basis of the information they are seeking.

5. CDC Coronavirus Chatbot

The chatbot powered by the Centers for Disease Control and Prevention, has also launched a chatbot on its official website to help the visitors do a basic self-assessment check on themselves by answering simple prompts and questions.

It is an example of how it is not just the tech-based companies that are making use of such digital measures but also government agencies and public health bodies.

With this we have looked into the different ways chatbots help with fighting the war against coronavirus rise along with the examples of some of the famous chatbots and companies behind them. The only thing left to do now is to get in touch with a software development agency that excels in chatbot creation or simply find ways to integrate existing chatbot systems in your business’s digital face. It will help you present yourself as a brand that cares about its customers and stakeholders’ wellbeing.



XBRL: scrapping quarterlies, explaining AI and low latency reporting




Here is our pick of the 3 most important XBRL news stories this week.

1 FDIC considers scrapping quarterly bank reports

The Federal Deposit Insurance Corp. is moving to boost the way it monitors for risks at thousands of U.S. banks, potentially scrapping quarterly reports that have been a fixture of oversight for more than 150 years yet often contain stale data.

The FDIC has been one of the cheerleaders and case studies for the efficiency increasing impact of XBRL based reporting forever. Therefore it will be fascinating to observe this competition and its outcome.

2 XBRL data feeds explainable AI models

Amongst several fascinating presentations at the Eurofiling Innovation Day this week was an interesting demonstration on how XBRL reports can be used as the basis of explainable AI for bankruptcy prediction.

The black box nature of many AI models is one biggest issues of applying AI in regulated environments, where causal linkages are the bedrock of litigation etc. Making them explainable would remove a major headache for lots of use cases.

3 Low latency earnings press release data

Standardized financials from Earnings Press Release and 8-Ks are now available via the Calcbench API minutes after published.  Calcbench is leveraging our expertise in XBRL to get many of the numbers from the Income Statement, Balance Sheet and Statement of Cash Flows from the earnings press release or 8-K.  

The time lag between the publication of earnings information and its availability in the XBRL format continues to be a roadblock for the wholesale adoption of XBRL by financial markets until regulators require immediate publication in the XBRL format in real time. The Calcbench API is a welcome stop gap measure. 



Christian Dreyer CFA is well known in Swiss Fintech circles as an expert in XBRL and financial reporting for investors.

 We have a self-imposed constraint of 3 news stories each week because we serve busy senior leaders in Fintech who need just enough information to get on with their job.

 For context on XBRL please read this introduction to our XBRL Week in 2016 and read articles tagged XBRL in our archives. 

 New readers can read 3 free articles.  To  become a member with full access to all that Daily Fintech offers,  the cost is just USD 143 a year (= USD 0.39 per day or USD 2.75 per week). For less than one cup of coffee you get a week full of caffeine for the mind.


Continue Reading


AI- hot water for insurance incumbents, or a relaxing spa?





The parable of the frog in the boiling water is well known- you know, if you put a frog into boiling water it will immediately jump out, but if you put the frog into tepid water and gradually increase the temperature of the water it will slowly boil to death.  It’s not true but it is a clever lede into the artificial intelligence evolution within insurance.  Are there insurance ‘frogs’ in danger of tepid water turning hot, and are there frogs suffering from FOHW (fear of hot water?)

image source

Patrick Kelahan is a CX, engineering & insurance consultant, working with Insurers, Attorneys & Owners in his day job. He also serves the insurance and Fintech world as the ‘Insurance Elephant’.

The frog and boiling water example is intuitive- stark change is noticed, gradual change not so much.  It’s like Ernest Hemmingway’s quotation in “The Sun Also Rises”- “How did you go bankrupt?  Gradually, and then suddenly!”  In each of the examples the message is similar- adverse change is not always abrupt, but failure to notice or react to changing conditions can lead to a worst-case scenario.  As such with insurance innovation.

A recent interview in The Telegraph by Michael Dwyer of Peter Cullum, non-executive Director of Global Risk Partners (and certainly one with a CV that qualifies him as a knowing authority), provided this view:

“Insurance is one business that is all about data. It’s about numbers. It’s about the algorithms. Quite frankly, in 10 years’ time, I predict that 70pc or 80pc of all underwriters will be redundant because it will be machine driven.

“We don’t need smart people to make what I’d regard as judgmental decisions because the data will make the decision for you.”

A clever insurance innovation colleague, Craig Polley, recently posed Peter’s insurance scenario for discussion and the topic generated lively debate- will underwriting become machine driven, or is there an overarching need for human intuition?  I’m not brave enough to serve as arbiter of the discussion, but the chord Craig’s question struck leads to the broader point- is the insurance industry sitting in that tepid water now, and are the flames of AI potentially leading to par boiling?

I offered a thought recently to an AI advocate looking for some insight into how the concept is embraced by insurance organizations.  In considering the fundamentals of insurance, I recounted that insurance as a product thrives best in environments where risk can be understood, predicted, and priced across populations with widely varied individual risk exposures as best determined by risk experience within the population or application of risk indicators.  Blah, blah, blah. Insurance is a long-standing principle of sharing of the ultimate cost of risk where no one participant is unduly at a disadvantage, and no one party is at a financial advantage- it is a balance of cost and probability.

Underwriting has been built on a model of proxy information, on the law of large numbers, of historical performance, of significant populations and statistical sampling.  There is not much new in that description, but what if the dynamic is changed, to an environment where the understanding of risk factors is not retrospective, but prospective?

Take commercial motor insurance for example.  Reasonably expensive, plenty of human involvement in underwriting, high maximum loss outcomes for occurrences.  Internal data are the primary source of rating the book of business.  There are, however,  new approaches being made in the industry that supplant traditional internal or proxy data with robust analysis of external data.  Luminant Analytics is an example of a firm that leverages AI in providing not only provide predictive models for motor line loss frequency and severity trends, but also analytics that help companies expanding into new markets, where historical loss data is unavailable.  Traditional underwriting has remained a solid approach, but is it now akin to turning the heat up on the industry frog?

The COVID-19 environment has by default prompted a dramatic increase in virtual claim handling techniques, changing what was not too long ago verboten- waiver of inspection on higher value claims, or acceptance of third party estimates in lieu of measure by the inch adjuster work.  Yes, there will be severity hangovers and spikes in supplements, but carriers will find expediency trumps detail- as long as the customer is accepting of the change in methods.  If we consider the recent announcement by US P&C carrier Allstate of significant staff layoffs as an indicator of the inroads of virtual efforts then there seemingly is hope for that figurative frog.

Elsewhere it was announced that the All England Club has not had its Wimbledon event cancellation cover renewed for 2021 (please recall that the Club was prescient in having cancellation cover in force that included pandemic benefits).  The prior policy’s underwriters are apparently reluctant to shell out another potential $140 million with a recurrence of a pandemic, but are there other approaches to pandemic cover?  The consortium of underwriting firms devised the cover seventeen years ago; can the cover for a marquee event benefit from AI methodology that simply didn’t exist in 2003?  It’s apparent the ask for cover for the 2021 event attracted knowledgeable frogs that knew to jump out of hot water, but what if the exposure burner is turned down through better understanding of the breadth of data affecting the risk, that there is involvement of capital markets in diversifying the risk perhaps across many unique events’ outcomes and alternative risk financing, and leveraging of underwriting tools that are supported by AI and machine learning?  Will it be found in due time that the written rule that pandemics cannot be underwritten as a peril will have less validity because well placed application of data analysis has wrangled the risk exposure to a reasonable bet by an ILS fund?

There are more examples of AI’s promise but let us not forget that AI is not the magic solution to all insurance tasks.  Companies that invest in AI without a fitting use case simply are moving their frog to a different but jest as threatening a pot.  Companies that invest in innovation that cannot bridge their legacy system to meaningful outcomes because there is no API functionality are turning the heat up themselves.  Large scale innovation options that are coming to a twenty-year anniversary (think post Y2K) may have compounding legacy issues- old legacy and new legacy.

The insurance industry needs to consider not just individual instances of the gradual heat of change being applied.

What prevents the capital markets from applying AI methods (through design or purchase) in predicting or betting on risk outcomes?  The more comprehensive and accurate risk prediction methods become the more direct the path between customer and risk financing partner also becomes.  Insurance frogs need not fear the heat if there are fewer pots to work from, but no pots, no business.

The risk sharing/risk financing industry has evolved through application of available technology and tools, what’s to say AI does not become a double-edged sword for the insurance industry- a clever tool in the hands of insurers, or a clever tool in the hands of alternative financing that serves to cut away some of the insurers’ business?  If asked, Peter Cullum might opine that it’s not just underwriting that AI will affect, but any other aspect of insurance that AI can effectively influence.  Frogs beware.

You get three free articles on Daily Fintech; after that you will need to become a member for just US $143 per year ($0.39 per day) and get all our fresh content and archives and participate in our forum


Continue Reading


MIT takes down 80 Million Tiny Images data set due to racist and offensive content




Creators of the 80 Million Tiny Images data set from MIT and NYU took the collection offline this week, apologized, and asked other researchers to refrain from using the data set and delete any existing copies. The news was shared Monday in a letter by MIT professors Bill Freeman and Antonio Torralba and NYU professor Rob Fergus published on the MIT CSAIL website.

Introduced in 2006 and containing photos scraped from internet search engines, 80 Million Tiny Images was recently found to contain a range of racist, sexist, and otherwise offensive labels such as nearly 2,000 images labeled with the N-word, and labels like “rape suspect” and “child molester.” The data set also contained pornographic content like non-consensual photos taken up women’s skirts. Creators of the 79.3 million-image data set said it was too large and its 32 x 32 images too small, making visual inspection of the data set’s complete contents difficult. According to Google Scholar, 80 Million Tiny Images has been cited more 1,700 times.

Above: Offensive labels found in the 80 Million Tiny Images data set

“Biases, offensive and prejudicial images, and derogatory terminology alienates an important part of our community — precisely those that we are making efforts to include,” the professors wrote in a joint letter. “It also contributes to harmful biases in AI systems trained on such data. Additionally, the presence of such prejudicial images hurts efforts to foster a culture of inclusivity in the computer vision community. This is extremely unfortunate and runs counter to the values that we strive to uphold.”

The trio of professors say the data set’s shortcoming were brought to their attention by an analysis and audit published late last month (PDF) by University of Dublin Ph.D. student Abeba Birhane and Carnegie Mellon University Ph.D. student Vinay Prabhu. The authors say their assessment is the first known critique of 80 Million Tiny Images.

VB Transform 2020 Online – July 15-17. Join leading AI executives: Register for the free livestream.

Both the paper authors and the 80 Million Tiny Images creators say part of the problem comes from automated data collection and nouns from the WordNet data set for semantic hierarchy. Before the data set was taken offline, the coauthors suggested the creators of 80 Million Tiny Images do like ImageNet creators did and assess labels used in the people category of the data set. The paper finds that large-scale image data sets erode privacy and can have a disproportionately negative impact on women, racial and ethnic minorities, and communities at the margin of society.

Birhane and Prabhu assert that the computer vision community must begin having more conversations about the ethical use of large-scale image data sets now in part due to the growing availability of image-scraping tools and reverse image search technology. Citing previous work like the Excavating AI analysis of ImageNet, the analysis of large-scale image data sets shows that it’s not just a matter of data, but a matter of a culture in academia and industry that finds it acceptable to create large-scale data sets without the consent of participants “under the guise of anonymization.”

“We posit that the deeper problems are rooted in the wider structural traditions, incentives, and discourse of a field that treats ethical issues as an afterthought. A field where in the wild is often a euphemism for without consent. We are up against a system that has veritably mastered ethics shopping, ethics bluewashing, ethics lobbying, ethics dumping, and ethics shirking,” the paper states.

To create more ethical large-scale image data sets, Birhane and Prabhu suggest:

  • Blur the faces of people in data sets
  • Do not use Creative Commons licensed material
  • Collect imagery with clear consent from data set participants
  • Include a data set audit card with large-scale image data sets, akin to the model cards Google AI uses and the datasheets for data sets Microsoft Research proposed

The work incorporates Birhane’s previous work on relational ethics, which suggests that the creators of machine learning systems should begin their work by speaking with the people most affected by machine learning systems, and that concepts of bias, fairness, and justice are moving targets.

ImageNet was introduced at CVPR in 2009 and is widely considered important to the advancement of computer vision and machine learning. Whereas previously some of the largest data sets could be counted in the tens of thousands, ImageNet contains more than 14 million images. The ImageNet Large Scale Visual Recognition Challenge ran from 2010 to 2017 and led to the launch of a variety of startups like Clarifai and MetaMind, a company Salesforce acquired in 2017. According to Google Scholar, ImageNet has been cited nearly 17,000 times.

As part of a series of changes detailed in December 2019, ImageNet creators including lead author Jia Deng and Dr. Fei-Fei Li found that 1,593 of the 2,832 people categories in the data set potentially contain offensive labels, which they said they plan to remove.

“We indeed celebrate ImageNet’s achievement and recognize the creators’ efforts to grapple with some ethical questions. Nonetheless, ImageNet as well as other large image datasets remain troublesome,” the Birhane and Prabhu paper reads.


Continue Reading
Blockchain12 mins ago

Cardano Price Hits 2020 Top Following Network Upgrade – Will the Rally Sustain?

Blockchain26 mins ago

The Tide is Turning, No-Coiner Bill Burr Convinced Over Merits of Bitcoin

Blockchain28 mins ago

Chinese e-commerce giant Alibaba leads in filing the most number of blockchain-related patents.

Blockchain39 mins ago

Russian court rules stolen Bitcoin cannot be returned to the owner

Blockchain41 mins ago

Ripple Price Analysis: Things Looking Grim for XRP as Bears Attempt To Push Below 1900 SAT

Blockchain45 mins ago

Tesla Stock Surpasses $1,200 — Now 30% Higher Than Bitcoin Market Cap

Blockchain45 mins ago

Altcoin market analysis: Cardano sees 14% price surge

Blockchain48 mins ago

Fundamentally Strong: Bitcoin Hit These Highs Today Despite Stagnant Price Action

Business Insider54 mins ago

The No. 1-ranked tech analyst on Wall Street says these 6 stocks have potential for huge gains as they transform the sector

Blockchain56 mins ago

Altcoin market analysis: SNX/USD has surged 120% over the past two weeks

Business Insider59 mins ago

China warns the UK it will take ‘corresponding measures’ to stop millions of Hong Kong citizens taking refuge in Britain

Blockchain1 hour ago

Litecoin Price Analysis: Holding Above 36.8 Level That Can Hold

Blockchain1 hour ago

IOTA Breaks Out, Aims For Recent Highs

Blockchain1 hour ago

Blockchain Explorer to Educate Users With a Bitcoin Transaction Privacy Score

Blockchain1 hour ago

RippleNet gets new client, but where is the GBP-USD corridor?

Business Insider1 hour ago

A 22-year market vet explains why stocks are headed for a ‘massive reset’ as the economy struggles to recover from COVID-19 — and outlines why that will put mega-cap tech companies in serious danger

Big Data1 hour ago

PyTorch Multi-GPU Metrics Library and More in New PyTorch Lightning Release

Blockchain1 hour ago

The On-Chain Case for an Imminent Bitcoin Bull Market Just Gained Strength

Blockchain1 hour ago

Bitcoin SV, Monero, BAT Price Analysis: 02 July

Blockchain2 hours ago

Russia’s Blockchain Voting System Let Users Decrypt Results Before Count

Blockchain2 hours ago

Bitcoin and Ether Market Update July 2, 2020

Blockchain2 hours ago

3 snippets to begin your day: Bitcoin’s been busy, another crypto-ETP and more

Private Equity2 hours ago

Priveq closes SEK2.5bn Fund VI launched amid worst of coronavirus

Blockchain2 hours ago

GTA Online Is Bigger Than Ever, Let’s Review it in 2020

Blockchain2 hours ago

The Bank of Canada says a CBDC should focus on accessibility and inclusion.

Gaming2 hours ago

Evening Reading – July 1, 2020

Private Equity2 hours ago

Angelo Gordon surges to $1.5bn hard cap for third Europe RE fund, almost double size of Fund II

Blockchain2 hours ago

Cardano, IOTA, Dash Price Analysis: 02 July

Blockchain2 hours ago

U.S. Authorities Point Searchlight into Crypto’s Role in Trafficking

Blockchain2 hours ago

Analyst Expects Bitcoin Above $9.5K in Near-Term as Risk-On Sentiment Improves

IOT2 hours ago

Panavise Speedwheel #3DThursday #3DPrinting

Cannabis2 hours ago

Former NBA Star John Salley Joins Insurance Pro Daron Phillips To Offer Cannabis Coverage

Private Equity2 hours ago

Kennet Partners raises €223m for biggest ever fund in tie-up with Edmond de Rothschild

Cannabis2 hours ago

CA Media Report: Border Patrol Seizing Cash and Cannabis From Legal California Operators

Cannabis2 hours ago

Congressman Cohen Wishes To Investigate and Consider the Impeachment of Attorney General William P. Barr Includes Reference To “pretextual antitrust investigations against industries he disfavors”

BBC2 hours ago

One in six jobs to go as BBC cuts 450 staff from regional programmes

IOT3 hours ago

Spinwheel – fidget toy #3DThursday #3DPrinting

IOT3 hours ago

Tube Cutter with Peephole easy fit #3DThursday #3DPrinting

Cannabis3 hours ago

Is THC Most Important in Good Weed?

CovId193 hours ago

Mudslide at Myanmar jade mine kills more than 100 people