Of course, hallucinations are quite annoying in many ways. However, does it cost too much for companies to use GenAI regarding cybersecurity issues other than simply taking some time to 'filter' or check out the fact? What do you think? The below is just a simple thought.
A Possible Scenario: People might ask GenAI tools to find some packages for solving coding problems. The tools would recommend some packages that are hallucinated. People in question could 'develop' the recommended packages in a malicious way by adding some factors on purpose. Some GenAI tools now have new information and recommend others the 'upgraded' packages. submitted by /u/Express_Fan7016
[link] [comments]
http://dlvr.it/T814TD
Post Page Advertisement [Top]
Subscribe to:
Post Comments (Atom)
';
(function() {
var dsq = document.createElement('script'); dsq.type = 'text/javascript'; dsq.async = true;
dsq.src = '//' + disqus_shortname + '.disqus.com/embed.js';
(document.getElementsByTagName('head')[0] || document.getElementsByTagName('body')[0]).appendChild(dsq);
})();
No comments:
Post a Comment