File Download Area

Information about "naplan 2015 state report year 7 with answers.pdf"

  • Filesize: 1.07 MB
  • Uploaded: 25/09/2019 11:43:53
  • Status: Active

Free Educational Files Storage. Upload, share and manage your files for free. Upload your spreadsheets, documents, presentations, pdfs, archives and more. Keep them forever on this site, just simply drag and drop your files to begin uploading.

Download Urls

  • File Page Link
    https://www.edufileshare.com/df8e422eb2363cc9/naplan_2015_state_report_year_7_with_answers.pdf
  • HTML Code
    <a href="https://www.edufileshare.com/df8e422eb2363cc9/naplan_2015_state_report_year_7_with_answers.pdf" target="_blank" title="Download from edufileshare.com">Download naplan 2015 state report year 7 with answers.pdf from edufileshare.com</a>
  • Forum Code
    [url]https://www.edufileshare.com/df8e422eb2363cc9/naplan_2015_state_report_year_7_with_answers.pdf[/url]

[PDF] naplan 2015 state report year 7 with answers.pdf | Plain Text

NAPLAN 2015 State report: Year 7 For all Queensland schools



i Queensland Curriculum & Assessment Authority | Contents Preface .................................................................................................. ............. 1 Placing the tests in the assessm ent context ........................................... ........... 2 Marking and scoring the tests ................................................................. ........... 2 Marking the tests ...................................................................................................... 2 Calculating raw scores ............................................................................................. 2 Constructing scale scores ........................................................................................ 2 Using scale scores ................................................................................................... 3 Understanding the data Which reports? ........................................................................................ ........... 4 Using data to improve teachi ng and learning ..................................................... 5 Year 7 Writing Writing prompt .............. ........................................................................... ........... 7 Key messages................. ........................................................................ ........... 8 About the task............................................................... ................................. ........... 8 Performance ............................................................................................................. 8 References ............................................................................................................... 9 Writing task sample ......... ........................................................................ ......... 10 Year 7 Literacy Language conventions ............................................................................ ......... 13 Spelling — Results and item descriptions ........ ............................................. ......... 13 Spelling — Key messages............. .................................................. ....................... 14 Grammar and punctuation — Results and item descriptions ................................. 17 Grammar and punctuation — Key messages......................................................... 18 Reading ............. ...................................................................................... ......... 20 Results and item descriptions................................................................................. 20 Key messages ........................................................................................................ 21 Year 7 Numeracy Results and item descriptions................................................................................. 24 Key messages ........................................................................................................ 26

| 2015 NAPLAN: State report ii

1 Queensland Curriculum & Assessment Authority | Preface The purpose of the National Assessment Program is to collect information that governments, education authorities and schools can use to determine whether Australian students are reaching important educational goals. As part of that program, the Literacy and Numeracy tests are valuable sources of information about literacy a nd numeracy learning that can be used to inform educational policy and current educational practice. The National Assessment Program — Literacy and Numeracy (NAPLAN) tests are developed using the nationally agreed Statements of Learning for English and Statements of Learning for Mathematics, 2005 . These statements describe essential skills, knowledge, understandings and capabilities that all young Austra lians should have had the opportuni ty to acquire by the end of Years 3, 5, 7 and 9. From 2016, the tests will relate to the Australian Curriculum. The NAPLAN tests are designed to provide a na tionally comparable indication of student performance in Language conventions, Writing, Reading and Numeracy. The tests are designed to assess a student’s ability to demonstrate the following skills: • Language conventions: The test assesses the ability of students to independently recognise and use correct Standard Austra lian English grammar, punctuation and spelling in written contexts. • Writing: The test assesses the ability of students to convey thoughts, ideas and information through the independent construction of a wr itten text in Standard Australian English. • Reading: The test assesses the ability of studen ts to independently make meaning from written Standard Australian English texts, in cluding those with some visual elements. • Numeracy: The test assesses students’ knowledge of mathematics, their ability to apply that knowledge in context independently, and their ability to independently reason mathematically. This document reports the performance of Q ueensland students in Year 7 who sat the 2015 National Assessment Program — Lite racy and Numeracy (NAPLAN) tests. Who should use this report? NAPLAN: State report will help teachers, principals and other school pers onnel understand, interpret and use the student performance informat ion contained in the test reports. Class and school reports are supplied electronically on the secure section of the Queensland Curriculum and Assessment Authorit y (QCAA) website: https://naplan.qcaa.qld.e du.au/naplan/pages/login.jsp . These reports are accessible only with the scho ol’s Brief Identification Code (BIC) login and password. Individual student reports are distributed to schools as printed copies. Principals Principals can use this document to help interpret their school reports and to provide information to the school community on aspects of the tests. The document pr ovides information on how to access and interpret the online repor ts located on the QCAA’s website. Curriculum leaders, Heads of Departme nt and Heads of Special Education Services Queensland’s performance on each of the Literacy and Numeracy strands is provided in this document. Curriculum l eaders can use this information to interpret the class reports.

| 2015 NAPLAN: State report 2 Classroom teachers Classroom teachers can us e information such as the item desc riptors, state and national results and the commentaries provided in this report to interpret their class reports. Teachers can compare the performance of thei r students on a particular item with Australian results. For example, an item with a low fac ility rate may not necessarily indi cate a problem in teaching and learning. It may be that this wa s simply a difficult item for all students in this cohort across Australia. The results for such an item may pr ovide information about the learning challenges associated with that concept but should not necessarily be cause for concern. Parents/carers Parents can use the information in this document to interpret the results on their child’s report. They are also able to judge how their child perfo rmed when compared with the whole population of students. The item descriptors provide useful information about the scope of the tests. Pre-service teachers Pre-service teachers will find the information in the commentaries on overall student performance useful in gaining an understanding of what students know and ca n do in some areas of Literacy and Numeracy at Year 7. Placing the tests in the assessment context The NAPLAN tests are national instruments desi gned to contribute to a school’s assessment program and to inform the teaching and learning cy cle. It must be remembered, however, that the results from the 2015 NAPLAN tests represent on ly one aspect of a school’s assessment program. The results from a school’s formal and informal assessment of students should be consistent with the NAPLAN test results. Principals and teachers should keep in mind that these were pencil-and- paper, point-in-time, timed tests. If the test results are different from what was expected, consider the possible reasons. The re sults of the tests may indicate as pects of student performance that need further investigation within the cla ssroom using other forms of assessment. Marking and scoring the tests Marking the tests The tests are scored against nationally agreed mark ing guides. There are four guides, one for the writing task and one each for the open responses in reading, numeracy and spelling. These guides provide information on the acceptable forms of the correct answer. For the Numeracy tests, students may provide a co rrect response in different forms. Professional officers review these resu lts and decide how to score. Calculating raw scores The simplest calculation made in scoring the test s is the raw score — the number of questions answered correctly. All of the questions for the Language conventions, Writing, Reading and Numeracy tests are marked as either correct or incorrect. Constructing scale scores Raw scores have limited use. They enable the performance of students who have all completed

3 Queensland Curriculum & Assessment Authority | the same test at the same time to be placed in a rank order, but they do not provide information about the level of difficulty of the test no r the relative differences between students. To achieve this, raw scores are transferred to a comm on scale that reflects how difficult it was to achieve each score. The scale is comparable between year levels for each assessment area. An equating process is also carried out on each year ’s test to enable scores to be compared between years of testing. This might mean, for example, that a raw score of 20 on the Year 3 Reading test is transformed to a scale score of 354. This will also represent the same achievement for a student with the same scale score in Year 5, and for a st udent with the same scale score for Reading in a previous year. The single scale for all students in all year leve ls is centred on approximately 500. Scale scores also provide a basis for measuring and comparing students’ abilities across years of schooling, for example comparing a student’s result in Year 3 in 2013 and Year 5 in 2015. Using scale scores The scale score can be used to compare the resu lts of different students. Principals and teachers should take care when making comparisons between small groups of students. For groups of fewer than 10 students, differences may not be reliable, particularly small differences. The scales can be used to monitor the growth of groups of students over time. Principals and teachers should ensure that the compositions of the groups are th e same. This enables the school to evaluate special programs that may have been put in place.

| 2015 NAPLAN: State reportUnderstanding the data 4 Understanding the data Which reports? Th e NAPLAN National Summary Report and the NAPLAN National report provide nationally comparable data about student performance within the National Assessment Program. These data provide states and territories with informat ion about the achievement of their students in relation to their peers across the nation. T hese data are available from the ACARA website. This NAPLAN State report provides detailed information about student performance on each of the test items. It gi ves information about: • the Queensland performanc e on each of the items • the national performance on each item • the item descriptors • some commentary on the state results • some recommendations for teaching. Together, these publications pr ovide system-level information and are generally available. The NAPLAN School reports give information about a school’s performance in each year level tested. They provide a summary of year-level performance as well as performance by gender, language background and Indigenous status in the following fields: • distribution of scale scores • distribution of achievement bands • school and state means • participation of the group. The shading showing the range of performance for the middle 60% of Queensland students, together with the state mean, locates a school’s performance relative to that of the state. NAPLAN data Government systems Australian public Schools Teachers Analysis of systems data: • Systems planning • Trends Analysis of school data: • Range • Comparisons of student & state Analysis of class data: • Test results by – class – group response Teaching, learning and assessment including planned explicit teaching and feedback based on identified learning goals. National report School report Class report

5 Queensland Curriculum & Assessment Authority | The NAPLAN Class reports show the performance of each stu dent on every item. They show the items a student had correct, including the errors made in each strand with the exception of reading, where the answers are generally too long to record. The report also gives the: • scale scores for each student • bands for each student • percentage correct for each item fo r the class and state, and by gender. The NAPLAN school and class reports are availabl e to schools from the QCAA secure website. Using data to improve teaching and learning While the national and state reports provide the comparative data, it is the class reports that provide a school with the information that can be used to inform teaching and learning and to build capacity in schools. Analysis of the NAPLAN cl ass data, in particular the performance on each item, will provide teachers wi th information about the understandings a nd patterns of misunderstandings in student learning. An analysis of the distracters presented in mu ltiple-choice items and the answers to the constructed-response items, other than those for reading, is available through the SunLANDA data analysis tool. This is available on the QCAA website and is designed to help schools with their analyses of class and school results. These results should be placed in a context with other school-based as sessments. Looking at the performance on the items and then analysing the error patterns allows teachers and principals to make hypotheses about why grou ps of students make particular errors. Schools can: • compare the facility rates (percent age correct) of items to see if their performance is consistent with the national and state resu lts available in this document • look at the common errors made by their stude nts and compare them with the common errors made in the state (Only errors from Queensland students are availabl e. These are found in the item analyses that are part of SunLANDA.) • form hypotheses about why studen ts are making these errors, e.g. – How did students think about this aspect of curriculum? – What misunderstandings migh t these errors represent? – How might the structure of the test question have shaped the response? Using a combination of the N APLAN data, school data and prof essional judgment, teachers should then test these hypotheses to see whether they are valid or whether there is more to be thought about and investigated. Inte rpretation of these results allows teachers to make judgments about teaching approaches and curriculum. The professional conversations that are part of this process are the most effective and powerful way to use the data as they are the vehi cle for developing shared understandings.

| 2015 NAPLAN: State reportUnderstanding the data 6

7 Queensland Curriculum & Assessment Authority | Year 7 Writing Writing prompt Simply the best Choose: UÊ ÊÌ…iÊLiÃÌÊ“œÛˆi]Ê/6ÊÃ…œÜÊœÀÊ«iÀvœÀ“>˜ViÊ ÞœÕÊ…>ÛiÊÃii˜Ê œÀÊ UÊ Ì…iÊLiÃÌÊLœœŽÊÞœÕÊ…>ÛiÊÀi>`°Ê 7ÀˆÌiÊ̜ʫiÀÃÕ>`iÊ>ÊÀi>`iÀÊÌ…>ÌÊÌ…iÞÊÃ…œÕ`Ê ÃiiÊœÀÊÀi>`ÊÜ…>ÌÊÞœÕÊ…>ÛiÊV…œÃi˜ÊÌœÊÜÀˆÌiÊ >LœÕÌ°ÊÊ UÊ Ê Start with an introduction. ˜Êˆ˜ÌÀœ`ÕV̈œ˜ÊiÌÃÊ>ÊÀi>`iÀÊŽ˜œÜÊÜ…>ÌÊÞœÕÊ >ÀiÊ}œˆ˜}ÊÌœÊÜÀˆÌiÊ>LœÕÌ° UÊ Ê Write your opinion on the topic. ˆÛiÊÀi>Ãœ˜ÃÊvœÀÊÞœÕÀÊœ«ˆ˜ˆœ˜°Ê Ý«>ˆ˜ÊÞœÕÀÊ Ài>Ãœ˜Ã° UÊ Ê Finish with a conclusion. ÊVœ˜VÕÈœ˜ÊÃÕ“ÃÊÕ«ÊÞœÕÀÊÀi>Ãœ˜ÃÊÃœÊÌ…>ÌÊ >ÊÀi>`iÀʈÃÊVœ˜Ûˆ˜Vi`ÊœvÊÞœÕÀÊœ«ˆ˜ˆœ˜° Remember to: UÊ «>˜ÊÞœÕÀÊÜÀˆÌˆ˜} UÊ ÊÕÃiÊ«>À>}À>«…ÃÊ̜ʜÀ}>˜ˆÃiÊÞœÕÀʈ`i>à UÊ ÜÀˆÌiʈ˜ÊÃi˜Ìi˜Vià UÊ ÊV…œœÃiÊÞœÕÀÊÜœÀ`ÃÊV>ÀivՏÞÊÌœÊVœ˜Ûˆ˜ViÊ>Ê Ài>`iÀÊœvÊÞœÕÀÊœ«ˆ˜ˆœ˜ UÊ Ê«>ÞÊ>ÌÌi˜Ìˆœ˜ÊÌœÊÞœÕÀÊëiˆ˜}Ê>˜`Ê «Õ˜VÌÕ>̈œ˜ UÊ ÊV…iVŽÊ>˜`Êi`ˆÌÊÞœÕÀÊÜÀˆÌˆ˜}ÊÜʈÌʈÃÊVi>À° © ACARA 2015 YEAR 7 AND YEAR 9

| 2015 NAPLAN: State reportYear 7 Writing 8 Key messages About the task In 2015, the NAPLAN Writing test used two prom pts for the first time, one for Years 3 & 5, and another for Years 7 & 9. Besides this difference, th e test conditions and administration of the test remained the same as in previous years, i.e. te achers delivered the same spoken instructions and read the text aloud to students. Working independently, students had to plan, compose and edit a written response. Students were allowed five minute s to plan, 30 minutes to write their script, and a further five minutes to edit and complete the task. Three pages were provided for students to write a response. The 2015 prompt for Years 7 & 9 was entitled Simply the best. Students were asked, in the textual component of the prompt, to choose the best movie, TV show, performance seen, or book read. They were then asked to persuade a reader that they should see or read what was chosen. Additional information was provided in the text ual component of the prompt. This named the structural components, and further defined these elements, e.g. Start with an introduction. An introduction lets a reader know what you are going to write about. Other notes were also provided in relation to the conventions associated with this type of writing task. A series of silhouetted images of various cultural activities, books, a nd film shots surrounded the textual component of the prompt. As was the case in 2013 and 2014, th e prompt was relatively open-ended, allowing students to base their writing on a topic of their own choice within the persuasive genre. Markers for this Writing test were trained usi ng the national persuasive writing marker training package, delivered as part of AC ARA’s national assessment progra m. Markers were recruited and trained in accordance with national protocols, app lied consistently across all states and territories. Registered teachers mark the NAPLAN Writing test in Queensland. All markers applied the 10 criteria and related standards from the marking rubric. Writing test scripts were marked on screen in all states and territories. Stringent quality-co ntrol measures were applied to the marking of student scripts, including a prescribed percenta ge of scripts to be double-marked, and the daily application nationally of control scripts for all markers. As part of the Queensland marking operation for 2015, referee marki ng was expanded to further ensure marking reliability. There is also provision for appeal over individual Writing test scores, once test results are released. On appeal, a student’s script is re-marked independently by two senior Writing test markers. The NAPLAN Persuasive writing marking guide is available at www.nap.edu.au/NAPLAN/About_each _domain/Writing/index.html. Performance In contrast to the performance of Years 3 and 5 st udents, there was little change in the Writing test performance for the older cohorts of Years 7 and 9. A possible prompt effect exhibited in Years 3 and 5 leading to improved performance was not rep licated in the upper grades. Other conditions (genre, time, test protocols) remained constant from 2014. The prompt did allow students to select a topic of personal interest with which they had some knowledge and familiarity. In fact, fe w students had difficulty in finding a subject on which their text could be based. Issues tended to emerge with re spect to the application of genre to the subject matter. So, students typically introduced their subj ect (e.g. favourite book or film) in an opening paragraph, stated broadly why the book/film/perfo rmance should be read/viewed, then proceeded to provide information regarding plot, characteri sation, cinematographic features etc. The danger here was that students were flirting with informativ e rather than the persuasive genre. Conclusions tended to focus on simple restatements of main points referred to in the body of the text. Often, there was a very close parallel between the word ing of the introduction and conclusion; on occasions, almost identical wording.

9 Queensland Curriculum & Assessment Authority | Some students, particularly Year 9 students, adopted a ‘review’ ty pe response, tending to provide information rather than persuasive argument defending their choice of book, film etc. Typically, this information was sandwiched between an intr oduction and conclusion which reflected some persuasive elements. So these st udent scripts were deemed to be ‘on genre’, though their final scores were impacted by the abs ence of persuasive elements used consistently throughout their texts. Because the Writing test is an ‘on de mand’ assessment, student responses frequently lacked detail about books, films and performances . So, a student may have been able to write with some fluency about generic features of their ch osen subject, but lacked substance in supporting the arguments provided. Exceptions to this we re cases where students clearly felt passionate about their topic, knew fi ne-grained details of the book or film, and were able to reflect this passion for the work in a clearly persuasive fashion. A small number of students adopted a text type that blended narrative and persuasive genres. Introductions provided something like a narrative or recount of an experience, leading in to the persuasive text proper. For inst ance, an introduction may have referred to an attendance at an evening performance of a play or musical, followed by a description of th e scene, any emotional response and so on. Features of the performance were then related in a persuasive body of text. There are risks in following this style, not the leas t of which is that the principal genre for the task may not be clearly developed, or not be outwardly apparent to the marker. On a more positive note, students in Years 7 and 9 showed that they are re ading novels, enjoying films, and attending other cultur al events. For many, this is an enjoyable experience on which they could readily reflect. Students generally wrote, on average, lengthier responses than in 2014, the prompt possibly having an effect in this regard. The 2015 test may also have suited students of a more academic persuasion, though students of varying ability levels were able to produce sound written responses on topics they were personally interested in. At this level, Years 7 and 9, students should be able to demonstrate control of the persuasive genre. The simple textual structure of introduction , body consisting of two to three paragraphs, and conclusion can sometimes inhibit the capacity of many students to more effectively persuade readers to their points of view . That is, the textual ‘form’ can become more dominant than the language purpose of actually persuading. Students who show greater command of the logical and textual features of persuasive writing, pa rticularly how this might play out in the body of the text, where persuasive tools such as condition, causation, comparison and supporting evidence are employed effectively, are rewarded through the NAPLAN marking rubric and in persuasive writing more generally. References Australian Curriculum, Assessment and Reporting Authority 2013, Australian Curriculum: English www.australianc urriculum.edu.au Queensland Curriculum and Assessment Authority 2013, Hidden worlds www.qcaa.qld.edu.au/downloads/p_10/3579_wt_hidden_worlds.pdf Queensland Curriculum and Assessment Authority 2011, Queensland’s Literacy Test: A framework for describing spelling items www.qcaa.qld.edu.au/downloads/p _10/3579_describing_spell_items.pdf

| 2015 NAPLAN: State reportYear 7 Writing 10 Writing task sample Year 7 — Glass Menagerie

11 Queensland Curriculum & Assessment Authority |

| 2015 NAPLAN: State reportYear 7 Writing 12 Year 7 — Glass Menagerie Audience 6 While this is not typical of Year 7 scripts, it shows that a high standard is possible. The student writes in the authentic style of a theatre review to persuade the reader of the high quality of this production and the play itself. By convention, a glowing review implies a recommendation to\ go. The strong sense of the reviewer’s personal feelings is an expected feature of this text type. Text structure 4 The introduction reveals the purpose.The statement about the play being astounding is not specific enough to be a main idea. The sections within the body are given relatively unhelpful signpost words Firstly, In addition and Furthermore . The actual sequence of topics found in the script is as follows: the acting; the audience response; the story and themes; and the design and its contribution to the play’s theme. The text should\ be tightened by making each subtopic relate to a main idea that is stronger and more specific. The brief but well-crafted conclusion includes the idea that the production breathed new meaning into the play. This would have functioned very well as a better main idea to unify the text. Ideas 5 The student discusses the relevant elements of a theatre production. Her judgments about t\ he quality of these elements are, in part, well explained and elaborated. Not all assertions are adequately backed up. For example, the assertion in the conclusion that the production was fragile suggests, but fails to explain, a link between the glass symbolism in the play and the production’s design. Persuasive devices 4 The writer’s personal appreciation of the play is for egrounded and this is an effective device. Her precise and specialist language (e.g. fractured characters, unhealthy dynamic ) makes it evident that the writer’s opinion is expert and well-considered. The rhetorically structured sentences maximise the force of the ideas they contain. Vocabulary 5 Despite the precise and sophisticated vocabulary, there are some clichés (e.g. hauntingly). There are some misuses which show where a teacher could help the student to improve, e.g. the highlight of everything is a confused idiom; content should refer only to the play’s subject matter, not to the lighting and costume design as well; the phrase truly stays (in the first sentence) lacks substance because the contrary idea of ‘false staying’ is not cogent. The student says the story and the other features are simply astounding (third sentence) but she actually means that the high quality of these features is astounding. Cohesion 4 The script uses vocabulary and grammar to keep a close focus on topics and subtopics. To a lesser extent, the script establishes relationships between the topics and subtopics. Paragraphing 3 Paragraphing enhances the flow of ideas. Each paragraph has a well-elaborated thesis \ sentence. Sentence structure 6 Sentences maintain a reflective and analytic tone and convey complex ide\ as and relationships. The student needs more control of prepositions, e.g. is comprised of should be ‘is composed of’ or ‘comprises’; contrasts against should be ‘with’; by La Boite Theatre should be ‘by the La Boite Theatre Company’. Punctuation 5 Punctuation is diverse and useful, e.g. phrasal adjectives with hyphens ( once-beautiful). Spelling 6 There are no errors and many words indicating mastery of derivational spelling (e.g. eponymous).

13 Queensland Curriculum & Assessment Authority | Year 7 Literacy Language conventions Spelling — Results and item descriptions The percentage columns give the proportion of correct answers (facility rate). These results are based on provisional data. Item Answer Qld% Aust% Description Proofreading — error identified 1 highway (hiway) 90 90.1 Correctly spells a compound word with a long vowel i spelled - igh. 2 shoulder (sholder) 85.9 86.4 Correctly spells a word with the long o vowel spelled - ou. 3 traditional (traditionel) 80.2 78.7 Correctly spells a word with the suffix - al. 4 onion (onyon) 75.7 78.1 Correctly spells a word with the consonant y represented by the vowel - i. 5prey (pray) 78.5 79.9 Correctly spells a homophone with the long a vowel spelled - ey . 6crate (craite) 75.8 76.6 Correctly spells a one-syllable word with the long a vowel spelled a-e. 7 pumpkin (pumkin) 72.4 75.3 Correctly spells a two-syllable word with the consonant p at the closed syllable juncture. 8 bought (bort) 75.3 74.5 Correctly spells a one-syllable word with the or diphthong spelled - ough. 9 penguin (pengwin) 67.1 71.1 Correctly spells a two-syllable word with the consonant w represented by the vowel u. 10 weighed (wayed) 66.6 67.7 Correctly spells a homophone with the long a vowel spelled - eigh . 11 honeycomb (honeycoam) 50.4 56.2 Correctly spells a word with the final silent letter b. 12 lifeguards (lifegards) 64.7 64.4 Correctly spells a word with the hard g consonant spelled - gu. 13 orbit (orbet) 61.3 58.4 Correctly spells a word with the schwa ( i) in the final unaccented syllable. 14 precipitation (precipitasion) 52.3 49.4 Correctly spells a word where the suffix - ion creates consonant alternation (precipitate/precipitation). 15 salmon (sammon) 46.8 52.7 Correctly spells a two-syllable word with the silent letter - l at the syllable juncture. 16 queue (quewe) 30.2 29.5 Correctly spells a one-syllable word with the long u vowel pattern - eue.

| 2015 NAPLAN: State reportYear 7 Literacy 14 Spelling — Key messages Performance For the most part the Quee nsland facility rates were very close to the national results. Words like traditional, bought, orbit, predictable, noxious and precipitation had slightly higher results. However, the Queensland performance was notably lower for the words honeycomb, salmon, penguin and helicopter . Items that required knowledge of how suffixes are added to stems or base words had notably low facility rates — foliage, noxious, successful and heavily. The low facility rate for these words is partially accounted for by the fact that students had first to find the error and then correct it. In the case of the first three words, students chose and misspelt other words in significant numbers so interpreting the results for these words needs to take that into account. Teachers should check their class reports of SunLANDA to see what their students did. Proofreading — error unidentified 17 predictable (predictible) 67.8 66.1 Identifies a mistake then correctly spells a word with the suffix -able. 18 helicopter (helecopter) 48.2 52.7 Identifies a mistake then correctly spells a word with the etymological element heli-. 19 experience (expirience) 46.7 46.6 Identifies a mistake then correctly spells a multisyllable word with an r-controlled vowel ( er). 20 interfere (intafere) 41.9 41.9 Identifies a mistake, then correctly spells a word with the prefix inter-. 21 categories (catagories ) 26.2 26.7 Identifies a mistake, then correctly spells a word in which the prefix cata- is spelled cate-. 22 successful (successfull) 46.8 45.8 Identifies a mistake then correctly spells a word with the suffix -ful. 23 heavily (heavally) 30.6 32.2 Identifies a mistake then correctly spells a word with the suffix -ly, requiring a change to the base word (change y to i ). 24 achieved (acheived) 37.5 39 Identifies a mistake, then correctly spells a word with the long vowel digraph - ie. 25 explanation (explaination) 44.4 45.7 Identifies a mistake, then correctly spells a word with the element -ation creating a vowel alternation with spelling change (explain/explanation). 26 relevant (relavent) 22.9 23.3 Identifies a mistake then correctly spells a word with a schwa (e) in the middle and the - ant form of the suffix - ent/- ant. 27 noxious (noxous) 26.7 26.6 Identifies a mistake then correctly spells a word where xi represents the sh sound. 28 contradicted (contrerdicted) 20 18.4 Identifies a mistake then correctly spells a word with the prefix contra -. 29 foliage (folige) 12.7 14.6 Identifies a mistake then correctly spells a word with the suffix -age. 30 maritime (maratime) 6.4 8.9 Identifies a mistake then correctly spells a word derived from the Latin mare, the sea (marine/marinate/maritime). Item Answer Qld% Aust% Description

15 Queensland Curriculum & Assessment Authority | In the foliage item, significant numbe rs of students thought dense and canopy were incorrect. The challenge for this word was not so much in the spelling of the suffix -age but in the spelling of the middle syllable where students we re given the correct spelling fo-li-age but which many changed in their attempts at correction, e.g. folage (11% of students made this error), or foilage (4% made this error). This suggests that stud ents find the middle syllable more difficult to spell than the suffix. Precise pronunciation will help stud ents to identify and spell the syllables in this word. Spelling the unstressed middle syllable is a ge neral problem evident in the spelling of other words such as categories, contradict, maritime, relevant. It is one of the major challenges for Year 7 students. Similarly, students chose excessive and awareness rather than the target word successful. Spelling this word correctly r equires students to add the suffix -ful to the base word without any change. They need to know that the suffix is spelt -ful. Many students who spelt this word incorrectly knew this. They correc ted the spelling of the suffix but introduced another error into the word. Typically the error was to drop one of the double letters in the base word, success. Students also did this in misspelling other words. This re presents the second spelling challenge facing Year 7 students — understanding syll able boundaries. They need to understand the common syllable patterns, e.g. vc/cv, bot h with the doublet as in suc/cess or where the consonants are different as in pump/kin or win/dow . They need to understand the role t hat the vowel plays in open and closed syllables. The vc/cv doublet is usually the first syllable pa ttern that students master. It is particularly important that they understand how and why thes e patterns work when they come to the NAPLAN spelling items. In these items, words with doublets are often used both as target words and distracter words in the sentenc es. When students don’t understand how the syllables work, they are tempted to randomly either add a letter as they did to misspell cannopy, or to delete a letter, as they did in misspelling successful or heavily. The third group of words that challenged Year 7 students were those where students needed to add a suffix.The conventions for adding inflectional endings that students learnt in the early years are used for adding affixes. Students would benef it from direct teaching of how to apply these principles. For example, 30.6% of Queensland students, 32.2% nationally, were able to correct and spell heavily. Almost all the students who were una ble to spell it correctly identified heavily as the misspe lt word but failed to realise it was a base word+suffix and that they needed to change y to i and then add the adverb-forming suffix - ly. Instead, students attempted to ‘sound-it-out’, producing errors like heavilly (14%), heavaly (10%), heavelly (7%) or heavely (5%). Student errors reveal that students often abandon advanced spelling strategies and revert to matching sounds to letters. Other aspects of suffixes that students need to know include the link between words ending in - ent and - ant and the suffixes -ence and -ance . Unfortunately, in the case of words ending in ent/ant, such as relevant, students need to just learn the word, paying attention to the syllables and articulating clearly as they do so. The word relevant had two errors, one in the middle syllable and the other with the -ant ending. Adding the adjectival form to - ible/-able — ibly/-ably — is a similar link that students should know, Students across Australia found the word noxious difficult to spell. Stu dents were drawn to other words such as conservation and eradication as the error. However, when they did recognise noxious as the word containing the error, they failed to recognise the error. The suffix is often preceded by a soft blend such as ci, ti and more rarely xi which combine to represent the sh sound, e.g . atrocious, delicious, ambitious and gi as in religious . Students can use the signpost of the -ous suffix as an indication to check the blend preceding the suffix. Words that required knowledge of Greek and Latin prefixes such as maritime, contradicted, categories and even helicopter also presented difficulties as st udents reverted to basic sound-it- out strategies. As student s develop their spelling abilities they ne ed to be able to spell using larger

| 2015 NAPLAN: State reportYear 7 Literacy 16 chunks of meaning or structure. For each of the previous words, the results would have been far higher if students had recognis ed the meaning units such as contra. Implications for teaching Test effects Often, students seem to allow themselves to be influenced by the printed misspelling of the target word instead of spelling the word using their normal practices. • To avoid the distracting effect of the printed misspelling, students should identify the target word, then cover up the misspellin g and then write the target word using their knowledge of spelling patterns and word components. • Standardised test results cannot substitute for assessments that assist teaching and learning. For teaching purposes, teachers should use measurements such as dictation tests, audits of student writing, and records of conferences with students. Strategies Teachers should teach stra tegies such as these: • Analysing syllables and stress patterns. Use this strategy for phonetically coded multisyllable words. • Applying knowledge of ‘rules’ for inflections and affixes. Use this for words containing grammatical components. • Applying knowledge of classical and foreign wo rd elements. Recall visual memories of the written form of mature vocabulary words. Use this for specialist and academic vocabulary. Content Spelling lessons should not be about random words bu t words that exemplify the separate layers of the spelling system and that suit secon dary students’ stage of learning these codes. Students should study: • mature vocabulary words with ‘within-word’ letter patterns. These include diphthongs, vowels in homophones and vowels in accented syllables (e.g. highway, shoulder, prey, crate, queue, weighed and achieved) • words with ‘syllable and affix’ spelling patter ns. For example, ‘open and closed’ syllable junction spelling is tested by the words orbit, onion, penguin, pumpkin, honeycomb, lifeguards and salmon • challenging verbs and tense inflections (e.g. br ing/bought) and suffixes attached to Latin elements (e.g. nox+ious, relev+ant, foli+age) and to base words (e.g. tradition+al, predict+able, success+ful). Students need to know that adding suffixes usually leaves the spelling of the stem unchanged (precipice to precipi-tate) but that there are exceptions such as explain/ explan ation, vigo ur/vig orous and pron ounce/pronunciation • words spelled according to ‘derivational rela tions’. Certain Greek and Latin elements recur and are spelled the same regardless of sound, e.g. maritime is related to marine, helicopter to helix , contradicted to contrary, categories to catalogue, interfere to interstate. Similarly, the tested word experi+ence is closely related to the word experi+ment and both carry the meaning of ‘from a test’. The per i- element sounds different in the tw o words but its spelling is constant because the meaning origin is constant.

17 Queensland Curriculum & Assessment Authority | Grammar and punctuation — Results and item descriptions The percentage columns give the relative pr oportion of correct answers (facility rates). These results are based on provisional data. Item Answer Qld% Aust% Description 31 D 93.8 94.1 Identifies a sentence requiring a question mark. 32 A 87 88 Identifies the correct verb tense to complete a complex sentence. 33 B 89.8 90.1 Identifies the correct use of a preposition. 34 D 87.7 86.5 Identifies the sentence containing a description (using adjectives). 35 B 85.6 86.7 Identifies an infinitive that completes a sentence. 36 D 79.8 80.1 Identifies the main clause that correctly completes a complex sentence. 37 B 80.6 79.8 Identifies correct punctuation of attributed direct speech. 38 C 78.9 79.2 Identifies the correct use of a comma after an introductory dependent clause. 39 C 79.9 80.5 Identifies the sentence that correctly combines information. 40 A 77.1 78.4 Identifies the correct use of brackets to enclose the explanation of an acronym. 41 C 78.4 78.3 Identifies the sentence that uses a contraction correctly. 42 B 60.4 61.8 Identifies a sentence containing an adjective. 43 A 73.1 73.8 Identifies the correct use of clause and list commas. 44 D 60.8 63.9 Identifies a command. 45 A 58.7 59.9 Identifies the sentence that uses correct capitalisation of proper nouns. 46 C 61 61.8 Identifies the incorrect use of a possessive apostrophe. 47 B 51.9 54 Identifies the correct synonym for a modal verb . 48 C 59.3 61.2 Identifies the correct suffix that changes a verb to an adjective. 49 B 51 49 Identifies the correct use of commas to separate words interrupting a main clause. 50 B 39 40.7 Identifies the correct use of hyphens for a compound adjective. 51 A 39.5 40.4 Identifies correct punctuation of attributed direct speech. 52 D 39.2 33 Identifies a compound sentence. 53 D 40.1 40.7 Identifies a sentence with correct verb tense. 54 B 29.6 32.5 Identifies the correct use of an adverb. 55 A 30.4 31 Identifies a sentence with a dangling modifier. 56 D 33.1 33 Identifies the grammatical functions of words in a context sentence. 57 B 26.2 27.2 Identifies the noun from a list of words with the same base. 58 D 14.4 14 Identifies the correct use of dashes to mark off a qualifying phrase.

| 2015 NAPLAN: State reportYear 7 Literacy 18 Grammar and punctuation — Key messages Performance Queensland Year 7 facility rates were higher than the national results for item 52 (identifying a compound sentence) item 34 (identifying a sentence containing a description) item 37 (punctuating direct speech) item 49 (on clause separation commas). Queensland performance was lower than the na tional one for most other items, notably: • item 44 (distinguish a command from a statement) • item 54 (avoid using an adjective as an adverb) • item 48 (recognise an adjective-forming suffix). Implications for teaching Formal and informal English The low facility on item 54 is a reminder that stu dents need to learn that standard English, which is used in formal contexts and in mo st writing, differs from informal English. Students may need explicit teaching of the differences. Grammatical terms Many questions on this test requ ired knowledge of the names of grammatical features (Items 44, 48, 53, 56 and 57). The low facility rates on these items suggest that many students need to learn the grammatical terms related to sentence modes, parts of speech , suffixes, prefixes and so on. Better knowledge of grammatical terms will allo w students to discuss language issues effectively. Teaching of formal grammar should be embedded in teaching functional communication. Students are likely to engage with grammar and punctuation if they can see that it helps them to understand and enjoy mature texts and to use the techniques of writers to improve their own writing. Assessment for teaching Teachers should assess and intervene when stude nts’ communication aims are being retarded or frustrated by their lack of gr ammar and punctuation knowledge. NAPLAN results cannot give teachers detailed asse ssment information. Instead, for baseline and post-lesson assessme nt, teachers could • audit student writing • audit student reading for fluency and comprehension • set exercises related to specific lessons. Content of lessons The teaching and assessing of gr ammar and punctuation should be • developmental (because it co vers increasingly mature skills) but timely (taught when students need to learn) and • systematic (covering the features relevant to th e levels of communication, from the whole text level to the sentence level and down to groups of words).

19 Queensland Curriculum & Assessment Authority | Comprehensive and specific information about what to teach from Years 1 to 9 is given in the draft scope and sequence for teaching grammar (and punctuation), available from the QCAA at: https://www.qcaa.qld.edu .au/downloads/p_10/qcar_ss_english_grammar.pdf Teachers can also refer to the following: 1. Books by Beverley Derewianka and Sally Humphr ey suggest ways of teaching that can often apply to older students. (Published by Prim ary English Teaching Association Australia). 2. Topics for teaching Year 7 grammar and punctuation are suggested in the Australian Curriculum English: • Recognise and understand that subordinate clauses embedded within noun groups/phrases are a common feature of writt en sentence structures and incr ease the density of information (ACELA1534) • Understand how modality is achieved through di scriminating choices in modal verbs, adverbs, adjectives and nouns (ACELA1536) • Analyse how point of view is generated in visual texts by means of choices, for example gaze, angle and social distance (ACELA1764) • Investigate vocabulary typical of extended and more academic texts and the role of abstract nouns, classification, descripti on and generalisation in building specialised knowledge through language (ACELA1537).

| 2015 NAPLAN: State reportYear 7 Literacy 20 Reading Results and item descriptions The percentage columns give the proporti on of correct answers (facility rates). These results are based on provisional data. Item Answer Qld% Aust% Description The pygmy 1 A 96.5 95.9 Infers a reason for a photograph. 2 B 89 88.9 Infers the significance of a map. 3 D 88.9 87.8 Interprets directly stated information. 4 C 91.6 91.1 Locates directly stated information. 5 B 95.1 94.5 Locates directly stated information. 6 A 79.2 81.1 Infers correct location of an additional fact. The mission 7 A 79.7 80.1 Infers meaning from a paragraph context. 8 A 83.5 82.6 Interprets the meaning of figurative language. 9 C 55.8 56.2 Interprets a character's response. 10 C 72.7 72.9 Interprets the meaning of a clause. 11 D 72.6 73.6 Identifies a change in a character's attitude. 12 C 39.8 39.3 Infers a character's role. The Minotaur 13 A 68.9 70.4 Infers a character's role. 14 D 33.6 34.4 Infers the meaning of a group's reaction. 15 C 29.1 29.4 Infers the background to an action. 16 B 37.3 36.7 Interprets the meaning of an expression. 17 B 56.8 56.6 Infers a character's motivation. 18 A 44.3 47.4 Identifies the specific purpose of brackets. The honey bee mystery 19 B 65.7 67.7 Locates directly stated information. 20 D 50.9 53.5 Integrates directly stated information. 21 D 46.4 47.2 Synthesises the main idea of a paragraph. 22 A 38.1 39.1 Identifies the purpose of a sentence. 23 C 67.2 66.6 Infers the reason for an action. 24 D 47.8 48.4 Infers meaning from an idiomatic phrase. 25 A 63.9 67.5 Infers the purpose of a concluding sentence.

21 Queensland Curriculum & Assessment Authority | * Item 30: Response identifies the sense of hearing AND one of the following: shrieked, grinding, tearing, swish, thump or crashing. Key messages Performance In 2015, Year 7 students were asked to complete 49 items as part of the Year 7 Reading test. Only one of the items, item 30, was not in multiple-cho ice form. For item 30, students were required to identify one of the five senses, and prov ide an example from the text provided, Hard running. Of the eight units in the test, two would be broadly identified as persuasive ( A way forward, Peer pressure ), three informative ( The pygmy marmoset, The honey bee mystery, Gliding through the deep ), and three imaginative ( The mission, The minotaur, Hard running ), though some of the stimulus texts, such as The minotaur, clearly crossed genres. It is a point worth noting that the Hard running 26 D 74.9 77.6 Identifies a source of tension between characters. 27 C 41.7 41.8 Infers the meaning of an idiom. 28 B 34.1 33.6 Infers the meaning of a word from context. 29 B 59.1 58.6 Infers the meaning of figurative language. 30 * 44.8 46.3 Locates and interprets vocabulary. 31 A 53.4 55.9 Interprets a character's motivation. A way forward 32 D 17.9 18.3 Interprets change in pronoun use. 33 B 55.7 56.1 Evaluates a character's stance. 34 A 41.7 42.7 Interprets the meaning of a phrase. 35 C 54.7 55.2 Infers the meaning of a sentence from context. 36 D 48.7 48.4 Synthesises information across a text. 37 B 37 36.9 Infers a perception of a character. Gliding through the deep 38 C 46 46.8 Interprets an explanation. 39 D 53.4 54.2 Interprets directly stated information. 40 D 24.5 25.5 Interprets the meaning of a statement. 41 A 47.7 49.8 Infers the purpose of a paragraph. 42 C 36.6 37.1 Integrates information to make an inference. 43 B 33.2 34.2 Synthesises information to make an inference. Peer pressure — a positive perspective 44 C 35.2 35.4 Synthesises information to make a judgment. 45 B 36.3 36.7 Identifies a synonym from context. 46 B 56.4 59.5 Identifies the main purpose of a paragraph. 47 A 34.8 35.8 Infers the meaning of a phrase from context. 48 C 28.4 29.1 Interprets information to make an inference. 49 B 50.1 50.2 Identifies a writer's tone. Item Answer Qld% Aust% Description

| 2015 NAPLAN: State reportYear 7 Literacy 22 hybrid generic text is commonpl ace in NAPLAN Reading tests. The performance of Queensland students in the Year 7 Reading test indicated improvement, both in terms of 2014 performance for Queensland students, and relative to the national mean score in the Year 7 Reading test, 2015. One pleasing as pect of this improvement was student performance in items based on narrative (imaginative) texts, with students performing near to, or above the national mean on 13 of the 18 items. Items base d on imaginative stimulus material have not previously been a strong point for Year 7 Queensland students. Overall, relative distance from the national mean on the Reading test items was small. There were only 6 items where the difference in Queensland student performance was more t han 2% lower than the national mean. These items, 18, 25, 26, 31, 41 and 46, did show an in teresting correlation in type, with the item descriptors for items 25, 41 and 46 requ iring students to identify a writer’s purpose in a sentence or paragraph, and items 26 and 31 based around character/s, either identifying a character’s motivation, or accounting for the tension existing between characters. In terms of gender, there has been a historical trend of girls out-performing boys in Year 7 Reading items, and this remained the case in 2015, though the gap appears to be narrowing, with quite small differences on many items. Girls still out performed boys on about 70% of the items. Some interesting results were noted on items 10, 20, 26 and 28, where gender differences were marked. Item 28, based on the narrative excerpt Hard running, asked students to find a synonym for the word onslaught . 44% of boys selected the correct option, assault, with only 23% of girls making this choice. Many girls selected the word disruption, instead, perhaps selecting the less ‘aggressive’ meaning. The nature of the readi ng and viewing materials to which students are exposed and the broader interests of boys and girls may also have played a role here. In item 20, on the other hand, 56% of girls se lected the correct response, compared to 47% of boys. Item 20, based on the information text The honey bee mystery, required students to unpack a set of information presented in a list. The task required logical analysis, steps of trial and error, and a degree of persistence. Th e latter skill is frequently evidenc ed in the performance of girls on NAPLAN Reading test items. Facility rates for items based on the final two units, Gliding through the deep (informative) and Peer pressure — a positive perspective were generally lower, Australia-wide, than other units in the Year 7 test for 2015. The performance of Quee nsland students reflected this trend, but results were generally aligned with national means. The di fficulty in items from these units was generated by a number of factors including the use of tech nical language, nominalisation, the inclusion of a number of main idea/purpose items requiring higher order evaluation, and length of texts. Queensland Year 7 students should be co mmended on their performance on these quite challenging texts and items located at the conclusion of the paper. Implications for teaching As has been the case in previous years, Queensl and Year 7 students have demonstrated capacity in literal comprehension. Extracting information from texts, the ‘right there’ type question is an area of some proficiency for Queensland students. It appears that simpler inferencing, particularly where the inference is predominantly drawn from the text itself, is also an improv ing skill for Year 7 students, as de monstrated by performance in the 2015 NAPLAN Reading test. Comprehension strategies directed towards higher-order comprehension still require additional attention in class and school programmi ng. These areas of comprehension include: • identifying the main idea or writer’s purpose in a text, or segment of a text • understanding what may constitute a character’s role, motivation or mood in a narrative text, and how this may change

23 Queensland Curriculum & Assessment Authority | • recognising how language can be used on many levels (figuratively) to enrich a text and encapsulate meanings or a writer’s intent • identifying strands or cohesive patterns in a te xt that contribute to main ideas, themes, tones etc. • developing understanding of the way in which the grammatical choices the writer makes shape textual form and therefore semantic relationships. Two recent papers located on the QCAA website deal with some of these areas, providing some specific strategies targeting higher order comprehension. https://www.qcaa.qld.edu .au/downloads/p_10/naplan_r ead_challenging_texts.pdf https://www.qcaa.qld. edu.au/downloads/p_10/napl an_use_reading_data.pdf

| 2015 NAPLAN: State reportYear 7 Numeracy 24 Year 7 Numeracy Results and item descriptions The numeracy strands are abbreviated as fo llows: Algebra, function and pattern (AFP); Measurement, chance and data (MCD); Number (N ); Space (S). All items are worth one score point. The percentage columns give the proporti on of correct answers (facility rates). These results are based on provisional data. Calculator-allowed paper Item Strand Answer Qld% Aust% Description 1 S B 80.8 80.7 Uses a scale to estimate a distance on a map. 2 S B 88.4 88.3 Identifies a solid with half the volume of a given solid. 3 N D 78.3 78.5 Calculates an amount using unit costs. 4 S B 72.9 73.3 Uses geometric properties to identify a 2-D shape. 5 N C 78.8 77.1 Determines the fraction of a collection. 6 MCD B 74.7 75.2 Uses a grid to compare the areas of triangles. 7 N C 74.1 74.9 Identifies a factor of a two-digit number. 8 MCD E 63.5 63.2 Interprets a histogram to find the frequency of several categories. 9 S A 77.7 77.4 Identifies the mirror image of an object. 10 N B 61 62.4 Expresses a quantity as a percentage. 11 AFP A 58.7 57 Identifies an equivalent expression that involves squares, multiplication and division. 12 AFP B 59.8 58.1 Identifies a rule connecting pairs of numbers in an additive pattern. 13 N A 46.8 48.9 Orders different representations of fractions in ascending order. 14 MCD D 48 47.6 Calculates a finishing time given the start time and duration. 15 AFP D 53.6 53.2 Identifies the number expression that represents a word problem. 16 S C 39.4 39.8 Calculates the size of an angle in an isosceles triangle. 17 MCD D 44.7 45.4 Calculates the probability of an event. 18 MCD C 51.4 52.1 Uses properties of a square to determine its side lengths. 19 N 5 062 043 47.2 47.4 Solves a multistep problem involving large numbers, addition and subtraction. 20 AFP B 63.9 64.1 Solves a problem involving an unknown number. 21 AFP 29 17.9 21.4 Solves a word problem involving division with a remainder.

25 Queensland Curriculum & Assessment Authority | Non-calculator paper 22 AFP A 40.1 39.1 Finds the value of an unknown in an equation involving simple fractions. 23 MCD D 27.4 30 Selects the numerical expression to calculate the area of a composite shape. 24 N C 39.5 40.4 Solves a word problem involving a rate — distance per hour. 25 MCD C 29.4 30.7 Interprets a table of data that involves overlapping categories. 26 S D 20.2 21.6 Solves a problem involving compass directions and measure of turn. 27 S 40 20.4 21.3 Calculates a length from a scale drawing. 28 N 350 13.3 16.2 Uses proportional reasoning to solve a problem involving two percentages. 29 AFP 12 13.9 16.7 Solves a multistep word problem involving the difference between whole numbers. 30 MCD B 34.8 34.9 Interprets a diagram and text to calculate the volume of a box. 31 N A 9.4 10.6 Determines the scale factor required for an enlargement. 32 S 120 6.7 8.2 Applies geometric properties of shapes and angles to calculate the size of an angle. Item Strand Answer Qld% Aust% Description 1 S C 83.5 84.1 Identifies a 3-D object from its net. 2 N B 86.6 86 Matches a diagram to the representation of a length that includes a fraction of a metre. 3 AFP A 87.4 88.3 Solves an equation with a missing value. 4 S D 80.8 80.9 Determines the number of hidden cubes in a 3-D stack. 5 N A 70.6 72.5 Solves a money problem involving multiplication and subtraction. 6 AFP B 76.1 75.7 Calculates a missing value in an additive pattern presented in a table of values. 7 MCD A 66.2 66.2 Uses estimation and reasoning to identify a right angle. 8 N 1000 67.1 69.1 Calculates a fraction of a quantity. 9 MCD A 61.3 62.1 Selects the most likely event from 2 spins of a spinner. 10 S B 71.5 72.8 Identifies the missing part of a symmetrical design. 11 S D 69.9 69.5 Identifies an incorrect face in a net. 12 N C 62.5 64.8 Subtracts four-digit numbers to approximate the age of an object. 13 MCD A 47 49.7 Interprets data in a pie graph. 14 N C 55 56.1 Calculates a percentage discount. Item Strand Answer Qld% Aust% Description

| 2015 NAPLAN: State reportYear 7 Numeracy 26 Key messages Performance Student results for Numeracy for Years 7 and 9 are reported as a single score. Where a student completes only one of the two Numeracy tests, thei r Numeracy score is an estimate of the score they may have received had they completed both tests. A significant difference between the raw scores achieved by an individual student on the two tests may warrant investigation. While the majority of students attempted to answer all test items, a significant number omitted the more difficult items towards the end of the test, pa rticularly items for which they were required to construct an answer rather than select a correct response from given options. These items are generally designed to differentiate student performance: to provide opportunities for higher performing student s to demonstrate their ability to solv e complex problems. The percentage of students failing to record an answer for construc ted-response items on the calculator-allowed test ranged from 4% to 16%. This was higher than the omit rates on the non-calculator test which ranged from 3% to 14%. While the omit rates are no t as high in Year 7 as in Year 9, teachers will need to ask students their reasons for failing to an swer questions, as a non-response provides no information for teachers to use to improve learning. 15 AFP C 50.2 50.7 Determines the value of a symbol in an informal equation. 16 MCD B 56.2 56.6 Identifies the expression to solve a measurement problem. 17 S D 49 47.8 Identifies a final compass point from given directions. 18 N B 48.6 49.7 Calculates a unit price to determine the best buy. 19 MCD C 42.1 43.4 Solves a problem involving measurement and proportional thinking. 20 AFP B 50.8 52.6 Identifies a future term in a repeating shape pattern. 21 MCD A 33.4 35.4 Solves a length problem. 22 N C 43.4 44.4 Estimates a cost using a rate of cents per minute. 23 AFP 6 28.3 30.7 Solves a problem involving sharing a quantity. 24 MCD E 18.6 20.2 Calculates the perimeter of a composite shape. 25 N 20 23.7 25.8 Solves a multistep problem involving a unit cost and total. 26 MCD D 22.1 24.7 Calculates the probability of an event involving factors. 27 S 3 18.7 21 Solves a problem involving scale, division and interpreting a diagram. 28 N 11; 13 11.9 15.5 Determines two odd numbers that are factors of a three-digit number. 29 MCD D 18.7 20 Interprets a timetable. 30 AFP D 14.9 16.2 Identifies equivalent number expressions. 31 S 6 11 11.6 Determines the number of faces of a prism. 32 N 18 3.7 5.4 Solves a multistep problem involving multiplication. Item Strand Answer Qld% Aust% Description

27 Queensland Curriculum & Assessment Authority | A number of items are common to both the Year 7 non-calculator test and the Year 5 Numeracy test. These ‘link’ it ems are for equating purposes and are intended to reflect differences between the two groups of students. Th e difference between the facility rates on these items should indicate that students at Year 7 have a significantly improved understanding of the relevant mathematical concept. There were 10 items that appeared on both the Year 7 non-calculator test and the Year 5 Numeracy test. Five of these (5, 7, 12,17, 27) were answered correct ly by 10% more of the Year 7 students than the Year 5 students. On two Space items (1, 4) the differences were 2.1% and 7.1%. Item 3, which required stud ents to solve an equation with a missing value, was answered correctly by 82.9% of Year 5 st udents and 87.4% of Year 7 student s. Approximately 9% more Year 7 students than Year 5 students were able to solve a problem involving measurement and proportional thinking (19). For item 24, in whic h students had to calculate the perimeter of a composite shape, Year 7 students outperform ed their Year 5 counterparts by 6.5%. On the calculator-allowed test, Y ear 7 male students outperformed female students on 21 of the 32 items with the difference being 5% or more on 5 items (10, 13, 21, 24, 28). Female students had a higher facility rate on 10 of the 32 it ems with the largest differe nce being 4% on items 5 and 11. The males also outperformed the female students on 19 of the 32 items on the non-calculator test. For 7 of these items, the difference was 5% or more with the greatest difference being 8% on items 8 (calculating a fraction of a quantity) , 14 (calculating a percentage discount) and 16 (identifying the expression to solve a measurement problem). On this test, female students outperformed male students on 9 items, with the greatest difference bei ng 5% on items 1 (identifying a 3-D object from its net) and 20 (identifying a future term in a repeating shape pattern). However, there is no pattern to these differences at the state level. The items on which one group has outperformed the other are not all from the same strand, do not test the same concepts or involve the same mathematical proc ess. It may be worth examining school and class data to see if there are any signif icant differences in your school. Implications for teaching Across the two tests, there were 19 items on which the facility rate for Queensland students was less than 30%. All these items occurred in the sec ond half of the tests, which means that they were intended to be more difficult than the items that preceded them. Grouping these items by the concept assessed sh ows 4 related to expressions, equivalence or equations; 4 related to rate, ratio, percent or propo rtion (all of which require multiplicative thinking); and 3 dealt with measurement — length, volume, mass area or time. Of the remaining items, two related to multiplication and division, while each of the others related to a different concept. Thirteen of the 64 test items were in some way related to the concepts of expressions, equivalence or equations. Not all were difficult and they included items in which students had to identify the expression that re presents a rule or a word problem (CA 12, CA 15, CA 23, NC 16), identify equivalent or non-equivalent numerical expressions (CA 11, NC 30), solve a numerical equation or calculate the value of an unknown in a numerical or symbolic equation (CA 22, NC 3, NC 15). The four remaining items (CA 20, CA 29, NC 23, NC 32), identified as being from the Algebra, function and pattern strand, are word problems for which students could have written an algebraic representation to help them calculate their answer. Almost two-thirds of the students were able to solve a problem involving an unknown number (CA 20). Whether they did this by wr iting the equation or by using a ‘guess and check’ strategy is something that teachers would need to ask studen ts. However, as problems become more difficult

| 2015 NAPLAN: State reportYear 7 Numeracy 28 with more variables, students will need to be able to use more efficient strategies. Often, the major stumbling block that stud ents experience when faced with a word problem, especially one that has a number of steps, is translating the problem into a number sentence or algebraic equation. To be able to represent a problem algebraically, students need to be able to recognise the relationships between the quantities or values in the problem. Teachers need to help students to do this by talking through a problem and explaining the thin king and reasoning involved in establishing the relationships. Going directly to writing equat ions means that students do not recognise the reasoning that needs to come before this. Teachi ng students how to use diagrams or tables to establish relationships will also help improve problem-solving sk ills. Begin by presenting students with a variety of problems, progressing from single- to multistep, and asking them to identify the steps needed to find a solution without the pressure of calculating it. Progress by asking students to develop the equations needed to solve problems, first in words then using mathematical symbols. Once again, begin with simple singl e-step problems and progress to more complex ones. Thinkboards are a useful way of having students represent problems in different ways and enable teachers to see a student’s level of unders tanding. Students should also be encouraged to estimate solutions to problems an d to explain the reason for this estimation. The ability to translate items CA 29, NC 23 and NC 32 into algebraic equ ations would have assisted students to solve these problems quite quickly and easily as the numbers involved are ‘friendly’. Of the 4 items on the tests with facility rates less than 30% that related to the concepts of rate, ratio, percent and proportion, three involved scale. In items CA 27 and NC 27, students were given diagrams that contained some of the relevant information while other important information was included in the text. As with all problems, t hese two items required careful reading. Students need to be taught how to annotate diagrams and shown ho w having all the relevant information in the one place aids problem solving. The other aspect of these two items and also of the third item involving scale — CA 31 — is that students nee d to understand what scale means and how to work with ratio. For item CA 31, students al so needed to understand the relationship between the area of a square and the length of its side. The fourth item in this set (CA 28) involved percentage and was answered correctly by only 13.3% of students . An earlier percentage item on this test (CA 10) had a facility rate of 61%. T he difficulty for students in this problem was recognising that the two given percentages were parts of the same whole — the target amount — and that the difference between these was the perc entage of the dollar amount given in the problem. To build students’ conceptual understandings of percentage: • use contexts that are familiar to students. Provide opportunities for students to engage with percentages • connect percentages to fractions — common and decimal — as this knowledge is useful for calculations, comparisons and checking needed in everyday situations (e.g. discounts, cost increases) greater than 100 and less than 1 • explicitly teach students to use th e percentage key on their calculator. Students found item NC 24, in which they were re quired to calculate the perimeter of a composite shape, the most difficult of the items related to l ength, volume, mass area or time. This item also appeared on the Year 5 test. The facility rate for Year 7 students was 18.6% and for Year 5 it was 12.1%. As this was a multiple-cho ice item, there was no likelihood of students confusing area and perimeter which may have been the case had it required a constructed response. The difficulty students encountered was identifying the miss ing measurements and including these in the calculation of perimeter. T hey need to be able to interpret all types of information graphics and to identify and calculate missing information. (Stu dents also needed this ability to interpret and annotate diagrams to answer item CA 32.) Item 23 on the calculator-allowed test, another mu ltiple-choice item involving a composite shape, also included a diagram. To answer this item correc tly, as well as knowing how to calculate the

29 Queensland Curriculum & Assessment Authority | area of a rectangle, students needed to ‘see’ that this shape could be divided into rectangles or extended to make a larger rectangle. Providing students with diagrams of composite shapes with some dimensions marked and asking them to calculate the ones that are missing may help develop their ability to interpret diagrams. Dividi ng composite shapes into simpler figures in order to calculate area may also be helpful. The style of the timetable in it em CA 23 may have been unfamiliar to the many students who answered this question incorrectly. It is importan t that students are exposed to a variety of timetables — both print and electronic — incl uding those that are not common in their geographical location. The other item related to the measurement of time was on the calculator- allowed test (CA 14). Students had to calculate a finishing time when given a start time and a duration. Fewer than 50% of students answered this item correctly with almost 30% of students using the conversion rate of 100 minutes in an hour and a further 23% of students failing to recognise that the duration extended from am to pm. Please refer to SunLANDA for a detailed analysis of individual test items, including teaching ideas designed to assist with the development of the understanding and skills required for each item. SunLANDA is available to all schools on the QCAA website. These materials are also available to Education Queensla nd schools through OneSchool . When looking at the data for a single test item , teachers can compare the grouped data for their class with that of the state or national coho rt. This will enable them to judge the level of difficulty that their students experienced with that item. For some items, the differences between the national, state and class data may not be signific ant, but teachers may wish to investigate the reasons for the poor performance of students on items that assess simple content and skills fundamental to numeracy development.