Skip to main content
 
 
 
IN THIS SECTION
ex. data visualization, research paper
  • What’s New with PatentsView — November 2024

    PatentsView has always been a leader in providing high-quality patents data to help drive insights into invention and innovation. The platform offers tools to help researchers better understand intellectual property (IP), inventors, and innovation. Users can also explore trends and connections between various topics to gain a deeper understanding of the IP landscape.

    Our team has been working diligently behind the scenes to not only uphold our reputation for high-quality data and disambiguation, but to make PatentsView better and more functional. Here are a few ways we are making PatentsView better for you.

    Service Desk

    We recently launched a new service desk to help users request an API key, get technical support, report a bug, or suggest improvements. The service desk also helps the PatentsView team better track requests and use your suggestions for continuous quality improvements.

    PatentSearch API

    The PatentSearch API’s full-text endpoints have been updated. For clarity and efficiency, granted and pre-grant text endpoints have been separated.

    • Granted text data can be retrieved at /api/v1/g_brf_sum_text/, etc. 
    • Pre-grant text data can be retrieved at /api/v1/pg_brf_sum_text/, etc.
    • The json keys for these endpoints' responses have been updated correspondingly to g_brf_sum_texts, etc.

    Full update notes can be found at https://search.patentsview.org/docs/2024/11/06/2.2-release.

    The new PatentSearch API is more advanced and efficient than the legacy API, which will be phased out in February 2025. Learn more about PatentSearch API in our PatentSearch API Reference page and Swagger interface.

    Ready to switch? Request a PatentSearch API key through our service desk.

    Ground Truth and Data Quality Checks

    To ensure the highest level of data accuracy, the PatentsView team has implemented several Ground Truth initiatives. These efforts involve cross-referencing patent data with verified sources to validate the information and correct any discrepancies. 

    By establishing a reliable ground truth, users can access more trustworthy data, which is crucial for conducting accurate analyses and making informed decisions. This commitment to data quality is what makes PatentsView the best in its class for patent data.
     

  • Support for Legacy API to End in February 2025. Switch to PatentSearch API Now.

    PatentsView is phasing out our legacy API, making way for the more advanced and efficient PatentSearch API. In September 2024, we began the process of retiring the old API, with full discontinuation set for February 2025. 

    We encourage all users to transition to the new PatentSearch API to ensure uninterrupted access to our services and to take advantage of the enhanced features and speed it offers.

    Ready to switch? Request a PatentSearch API key through our service desk.

    About PatentSearch API

    The new PatentSearch API is intended to inspire exploration and enhanced understanding of US intellectual property (IP) and innovation systems. The database for the PatentSearch API is updated regularly and features the best available tools for inventor disambiguation and data quality control.

    Researchers and developers can use the API to uncover information about people and companies, and to visualize trends and patterns in the US innovation landscape.

    The API offers seven unique endpoints that allow users to explore various questions, such as:

    • Which companies hold patents in 3D printing? Discover their locations and the technologies they were innovating in before and after receiving 3D printing patents.
    • What technology has been most commonly patented in the US in the last five years? Identify the top US and non-US cities producing these patents.
    • Which US inventors earned the most patents in the last 30 years? Track their patenting activity, including the number and types of patents and their co-inventors.

    Learn More About PatentSearch API

    For detailed documentation on the new PatentSearch API, visit our PatentSearch API Reference page. Additionally, you can explore the Swagger interface.

    Why Are We Switching APIs?

    PatentsView has been offering an API since 2015, which has been widely used and valued by thousands of users. 

    However, based on years of feedback and the evolving nature of patent data, PatentsView released the PatentSearch API in early 2024. This new API enhances the functionality and speed of the previous version.

    Consolidating Names

    Over time, the legacy API has been known by various names, including “PatentsView API,” “Swagger-based API,” and “MySQL API.” The new PatentSearch API has also been referred to as the “Elasticsearch API,” “Beta API,” and “Search API.” 

    Moving forward, we will consolidate the naming to its official title – PatentSearch API.

  • Do Patents Drive Investment in Software? A Spotlight on the Impact of Alice Corp. v. CLS Bank International

    In the world of software startups, patents are often seen as a golden ticket to securing investment. But do they really hold that much sway? A recent study by James Hicks looks into this question, focusing on the impact of a landmark Supreme Court decision, Alice Corp. v. CLS Bank International, on software patents and investments.

    The study found that patents on certain types of software did not necessarily increase investment and other positive business outcomes. The findings encourage us to look more closely at when patents are needed to protect intellectual property.

    What is the Alice decision?

    The Alice decision changed the criteria for software inventions to be eligible for patents. The case involved an electronic service that implemented financial trading systems. The Alice ruling stated that the service was based on an abstract idea, and therefore was not patentable. In practice, businesses applied this principle to any software involving business methods, which led to a dramatic drop in investment in this type of patent.

    Despite this, Hicks' study found no evidence that this drop in the number of patents led to less early-stage venture capital investment or affected outcomes such as acquisitions and initial public offerings (IPOs).

    The results indicated that patents may not be necessary for investment in the software industry. While patents are needed to protect certain types of intellectual property, these findings suggest that we may need to take another look at the value of patents in other areas.

    PatentsView Data: Unraveling the Relationship

    The study used data from PatentsView to identify companies that applied for patents and compare the number of patents with investment in the same company. This helped researchers understand the impact of the Alice decision on patent approval and subsequent investment in business-methods software firms. The results indicate that patents did not have a discernible effect on funding or successful outcomes in the software industry.

    Reevaluating the Role of Patents in Software Innovation

    The relationship between patents and investment in innovation is complex. This study suggests that the costs of obtaining patents may not always justify the benefits and invites more research into understanding when patents add value to a business. PatentsView data can help future researchers highlight connections and provide insight that will help inventors and businesses make better decisions about their intellectual property.

    Disclaimer: This post was written with the assistance of artificial intelligence.

  • Discovering value: women’s participation in university and commercial AI invention

    Artificial Intelligence (AI) is making waves in every corner of the world, and women are playing a significant role in this revolution, especially in the field of biotechnology AI, according to a recent study published in Nature. The study, titled Discovering value: women’s participation in university and commercial AI invention, used PatentsView data to identify the gender of inventors before using World Intellectual Property Organization categories to determine which patents contain AI. 

    It was written by Alexander Giczy, Nicholas Pairolero, and Andrew Toole.

    More Women are Participating in AI Innovation

    The study found that women are making significant contributions to AI innovation, but there is still work to be done to ensure that people from many different backgrounds are participating in the development of AI.

    The authors analyzed the women as inventors rate (WIR), which is a measure of the number and share of people receiving patents who are women. They found that the overall number of women inventors who received AI-related patents has increased from around 500 per year in 1995 to around 9,000 in 2020. The WIR for AI inventions increased from 10.1% to 12.6% in the same time frame.

    “These results show that women are not only participating in AI invention but are doing so at a slightly higher rate than in non-AI technologies,” the authors said in the study.

    They found differences in the AI WIR based on sector (e.g., AI and biotechnology), and whether the inventors worked for companies or universities.

    The Value of Women Inventors

    The authors noted research showing that diverse teams are more likely to improve innovation and business performance by looking at problems from different perspectives. Testing that theory, they found that teams with a higher proportion of women tended to be associated with higher economic value of granted patents. 

    “While such an analysis does not imply causation, the value of the patent is higher for teams with relatively more women: a patent with an equal number of male and female inventors has a value approximately $1.038 million higher than a patent with all-male inventors,” the authors said in the study.

    How PatentsView Can Help

    This study contributes to a growing body of research that shows how important diversity is for innovation and emerging technologies. To thrive and grow in the future, everyone must be able to participate in innovation and patenting, regardless of their background. Tools like PatentsView can help us track our progress and identify areas where we can improve.

    You can explore more on this topic on PatentsView’s Gender & Innovation and AI & Innovation pages.

  • What’s New with PatentsView – March 2024

    The PatentsView team is always working to make our data more complete, more accurate, and more useful. Recently, we completed a validation process of our assignee disambiguation algorithm. We created large, updated, ground truth dataset to calculate a current view of the performance of our assignee disambiguation algorithms.

    This blog highlights our continued commitment to assignee disambiguation validation by detailing the process we went through to build a hand-labeled ground truth dataset, the python packages and metrics used in assignee validation, and by openly disclosing results and their statistical significance. 

    Why is evaluation important?

    Model evaluation is important because it allows us to be publicly transparent about the performance of our disambiguated data, identify situations where our model does not perform as well as it could, and to make changes to improve overall performance.  We validate our assignee disambiguation algorithm by estimating key performance metrics, like precision and recall, that summarize its accuracy. 

    Assignee Labeling Process

    Estimating the performance of our disambiguation algorithm requires benchmarking data: some “ground truth” against which we can compare our predictions and assess the quality of our disambiguation. There are two main types of ground truth data used to evaluate entity resolution algorithms. 

    The first type is partially resolved entities, which consists of examples of matching entity mentions (e.g., an example of assignees from two patents, like “Microsoft” and “Microsoft Corp.” that we know refer to the same organization), as well as examples of non-matching entities (e.g., “Microcorp” and “Microsoft”), which upon research of company location and services offered, we know are two separate companies. 

    The second type of ground truth data is fully resolved entities. In this case, we find all patents for which Microsoft is an assignee and use that as complete ground truth for evaluation. We demonstrate how we cluster entities, such as all instances of “Microsoft,” to create our ground truth in the remaining paragraphs of this section.

    Our evaluation process focuses on the second type of data, fully resolved entities, because this method provides more robust statistical outputs. We employed three data labelers for over 100 combined hours to resolve the entities of over 250 randomly selected assignee mentions. To maximize the accuracy of the ground truth labels we created, that is groupings of rawassignees that are mentions of the same organization, we broke the process down into two main parts: (a) finding everything that could be a related entity and then (b) removing unlike assignees based on greater rawassignee detail.

    In step (a) finding everything that could be a related entity, for each assignee, we compared the assignee reference (organization name or first/last name) with hundreds of similar names in our database. This was done by the hand-labelers using a custom-made Streamlit application, which we designed to be both a query and data augmentation tool. Labelers pulled similar assignees by testing various potential name representations – name shortenings, abbreviations, potential misspellings, etc. – in Streamlit and saving the results. Streamlit then augmented the saved results from the labeler search by adding previous clusters (prior disambiguation result) that were associated with one or more of the rawassignees found by the labeler and were not previously included. 

    For step (b) removing unlike assignees based on greater rawassignee detail, hand-labelers reviewed the saved cluster output from step (a). The saved cluster data contained additional information about the rawassignee, including the associated patent, patent type, CPC classification, and location. Using filters, sorting, or any resource necessary, labelers carefully inspected all types of assignee data which could prove useful to remove any rawassignee mentions that should not be included in the final cluster. Dependent on the size of clusters, manual review could take between two minutes and an hour.

    Evaluation Packages – Written for PV

    PatentsView estimates performance metrics in a transparent manner using open-source entity resolution evaluation software. The functions are called in this location from our PatentsView-DB repository. We leverage two repositories; ER-Evaluation for the pairwise precision and recall metric functions and PatentsView-Evaluation to upload the relevant benchmark datasets (Binette & Reiter, 2023,1). You can find more technical details about this process in the documentation and code linked in this section.

    Statistical Significance

    Precision and recall are standard metrics in evaluating entity resolution and a detailed discussion about those metrics can be found in our last evaluation report (Monath, et al, 2021, 15). Based on the newest PatentsView Ground Truth labeled data, the latest data update achieved a precision of .90 and a recall of .72. This indicates that we are more likely to leave a rawassignee record[1] out of a cluster (False Negative) than erroneously include an additional record into a cluster (False Positive). 

    Precision and recall are calculated on an entity level evaluating the results of the most recent assignee disambiguation algorithm for the 228 assignee clusters, where we have ground truth data. See our last evaluation report for a more detailed explanation on the difference of entity-level versus rawassignee record level evaluation (Monath, 2021, 16-17). A standard deviation of around 4% for both of our estimates can be interpreted as 4% variability around the estimate – meaning that there is an approximately 68% likelihood that the true (population) precision is between 0.8604 to 0.9400, and that recall is between 0.6839 to 0.7699. This team believes that 4% variability is a narrow enough range for confidence in these evaluation metrics. 

    MetricEstimateStandard Deviation
    Precision0.9000.040
    Recall0.7270.043
    F10.804 

     

    Conclusion

    In conclusion, the recent advancements in PatentsView, particularly concerning the validation of assignee disambiguation algorithms, signify a steadfast commitment to data accuracy and transparency. Through meticulous evaluation processes and the creation of comprehensive ground truth datasets, we ensure quality of our disambiguated data.

    By employing multiple data labelers and leveraging sophisticated evaluation packages, such as ER-Evaluation and PatentsView-Evaluation, metrics like precision and recall are estimated, shedding light on the algorithm's performance.

    The latest update boasts an impressive precision of 0.90 and a recall of 0.727, indicating a high level of accuracy in entity resolution. These efforts underscore PatentsView's unwavering dedication to providing users with high-quality disambiguated assignee data and our commitment to transparency to our users in our processes and our work.

    Citations

    Binette, O., & Reiter, J. P. (2023). ER-Evaluation: End-to-End Evaluation of Entity Resolution Systems. The Journal of Open-Source Softwarehttps://joss.theoj.org/papers/10.21105/joss.05619.pdf

    Monath, N., Jones, C., & Madhavan, S. (2021, July). PatentsView: Disambiguating Inventors, Assignees, and Locations. Retrieved from https://s3.amazonaws.com/data.patentsview.org/documents/PatentsView_Disambiguation_Methods_Documentation.pdf 

     

    [1] PatentsView defined a “rawassignee record” as every mention of an assignee found on all granted patents and patent applications

Button sidebar