In Macros, How do I refer to entire scraped article from search engine?
I see you can pull %scraped_sentence% and %scraped_paragraph% … What if I need a larger context ?? … (which I do in at least one use case).
Is there a %scraped_article% macro possible? While I don’t plan to copy entire articles, I need the AI to be able to understand the entire article as RAG.
But it may be unrelated, right? Meaning each paragraph is randomly selected from different articles.
It’s hard to teach an AI based on random thoughts, especially if the aim of the articles are different in scope.
The “advantages of home gardening” vs “gardening tools from Home Depot”. etc
We are going to end up with AI articles that seem to have random tangent thoughts. I already had one talk about a whole different topic due to a homonym.
I was hoping I could feed the AI article by article.
How do I train the AI coherent data from the web. Am I missing something here?
Article by article may not work actually, since in one article you could have both “advantages of home gardening” vs “gardening tools from Home Depot” topics.
I suggest you try instead %scraped_title% or %scraped_subheading%
If what you want is a longform article, you should be using ai outline template.
Basically, what I’m trying to do, is train the AI with articles from the top X results … have AI craft an outline based on what it learned WHILE FOCUSING ON THE KEYWORD and AI-WRITTEN TITLE … and then complete the article based on RAG (Retrieval Augmented Generation) from the Top X articles. The point here is to prevent AI hallucination and provide reasonably-certain fact-based articles. However, it’s not easy to accomplish this when paragraphs are jumbled by random selective inclusion and order. See the dilemma?
The method you are proposing by creating the outlines etc, you are still subject to AI hallucination based on the model’s original training. This will always be a problem, unless the model is fine-tuned when creating the output. This fine-tuning and humanizing the writing style is also why the output takes so long and breaks the 180s timeout.
FYI, Many LLMs now have 128k context windows … so there is plenty of room to add one article at a time … and that is only going to grow.
Just want to chime in, since I’m going to use scrape_article for my use case.
I think scrape_article is very useful, it can be compressed using SPR* prompt for llm context (LCW vs RAG, long content window wins in every benchmark).
The main use of SPR (sparse priming representation) is for token efficiency (without losing original meaning to avoid hallucination/forced creativity).
So instead of using full article text as llm context, use SPR (a compressed version).
Unpredictable Tech, Predictable Humans: Human desires drive tech evolution; tech itself is stochastic. Core Human Desires: Safety, success, social connection (STC). AI Predictions: Future developments in AI based on human needs, not tech trends.
Digital Assistants (DAs):
Central to personal AI ecosystem.
Will know everything about users (health, finances, preferences).
Users will trade privacy for functionality.
API Proliferation:
Everything will have an API (Daemon).
Businesses and individuals will broadcast capabilities.
DAs will interact with these APIs for user benefit.
Mediation by DAs:
DAs will manage interactions with numerous APIs.
Users will express needs; DAs will fulfill them.
Active Advocacy:
DAs will monitor and filter information to protect users.
Continuous defense against manipulation and threats.
Ecosystem of DA Modules:
Companies will create specialized modules for DAs.
Marketplace for modules will thrive, enhancing DA capabilities.
AR Interfaces:
Augmented reality will visualize DA data.
Users will see character, personality, and functionality of entities around them.
Role-Based DAs:
Multiple DAs for specific tasks, each with distinct personalities.
Collaboration among DAs for holistic support.
Security Concerns:
Risks of DA hacks and API vulnerabilities.
Potential for influence operations through DAs.
Societal Implications:
Impact on human relationships and privacy.
Possible polarization in communication styles due to DA presence.
Conclusion:
New AI ecosystem driven by human needs is inevitable.
Focus on safety, thriving, and connection will shape tech development.
SPR … sounds great … Basically you’re asking the LLM to take an entire 3000 word article and give me “just the facts, maam”. I’m going to do some testing on this and let you know how it works using his model.