This code is running successfully but it is not taking information from knowledge pdfs.
from crewai import Agent, Crew, Process, Task
from crewai.project import CrewBase, agent, crew, task
from crewai.knowledge.source.pdf_knowledge_source import PDFKnowledgeSource
@CrewBase
class KbDemo():
agents_config = 'config/agents.yaml'
tasks_config = 'config/tasks.yaml'
pdf_knowledge_source = PDFKnowledgeSource(
file_paths=["updated_Resume.pdf","PSResume.pdf"]
)
@agent
def analyst(self) -> Agent:
return Agent(
config=self.agents_config['analyst'],
allow_delegation=True,
verbose=True,
)
@agent
def reviewer(self) -> Agent:
return Agent(
config=self.agents_config['reviewer'],
allow_delegation=False,
verbose=True,
)
@task
def analyst_task(self) -> Task:
return Task(
config=self.tasks_config['analyst_task'],
)
@task
def reviewer_task(self) -> Task:
return Task(
config=self.tasks_config['reviewer_task'],
output_file='report.md'
)
@crew
def crew(self) -> Crew:
return Crew(
agents=self.agents,
tasks=self.tasks,
process=Process.sequential,
verbose=True,
knowledge_sources=[self.pdf_knowledge_source],
)
------------------ Output ------------------
# This output is not related to my pdf , I am not sure from where it is getting information.
**Best Fits for Data Engineering Roles from Knowledge**
1. **Candidate Name**: John Doe
**Contact Details**: john.doe@email.com | (555) 123-4567
**Skills**:
- Programming Languages: Python, SQL
- Data Warehousing Experience: Amazon Redshift, Google BigQuery
- ETL Process Knowledge: Talend, Apache NiFi
- Big Data Technologies: Hadoop, Spark
2. **Candidate Name**: Jane Smith
**Contact Details**: jane.smith@email.com | (555) 987-6543
**Skills**:
- Programming Languages: Python, Java
- Data Warehousing Experience: Snowflake, Microsoft SQL Server
- ETL Process Knowledge: Apache Airflow, Informatica
- Big Data Technologies: Kafka, Spark
3. **Candidate Name**: Mike Johnson
**Contact Details**: mike.j@email.com | (555) 555-1212
**Skills**:
- Programming Languages: SQL, R
- Data Warehousing Experience: Oracle, Teradata
- ETL Process Knowledge: SSIS, Apache Beam
- Big Data Technologies: Hadoop, Flink
4. **Candidate Name**: Emily Davis
**Contact Details**: emily.d@email.com | (555) 246-1357
**Skills**:
- Programming Languages: Python, Scala
- Data Warehousing Experience: Redshift, Azure Synapse
- ETL Process Knowledge: Apache Airflow, Talend
- Big Data Technologies: Spark, Hive
5. **Candidate Name**: David Brown
**Contact Details**: david.b@email.com | (555) 333-4444
**Skills**:
- Programming Languages: Java, SQL
- Data Warehousing Experience: PostgreSQL
- ETL Process Knowledge: SSIS, Informatica
- Big Data Technologies: Spark, Kafka
**Recommendation**: All listed candidates exhibit strong programming fundamentals, substantial experience with data warehousing solutions, and proficient knowledge in ETL processes and big data technologies. Conducting interviews with these candidates is highly encouraged, as they align superbly with our Data Engineering role requirements and are likely to make significant contributions to our team's projects. Their varied skill sets embody the adaptability and expertise needed in our data engineering operations.